[ 507.835817] env[63197]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63197) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 507.836197] env[63197]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63197) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 507.836301] env[63197]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63197) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 507.836607] env[63197]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 507.931576] env[63197]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63197) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 507.941658] env[63197]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=63197) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 508.541974] env[63197]: INFO nova.virt.driver [None req-111d852c-d1ef-4406-9c86-0c7cfa0efbed None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 508.611740] env[63197]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.611891] env[63197]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.612027] env[63197]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63197) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 511.698168] env[63197]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-c6bf0752-ee1a-41e9-9da0-ce7112af1905 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.713834] env[63197]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63197) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 511.714246] env[63197]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-9f3f710b-7b3a-4b79-811c-001b1efeacdc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.749533] env[63197]: INFO oslo_vmware.api [-] Successfully established new session; session ID is d0ecf. [ 511.749698] env[63197]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.138s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 511.750221] env[63197]: INFO nova.virt.vmwareapi.driver [None req-111d852c-d1ef-4406-9c86-0c7cfa0efbed None None] VMware vCenter version: 7.0.3 [ 511.753530] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83fa8cf-b9f6-4eee-a6a3-5581cd7bdf98 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.770487] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5653fa-3677-4dde-b3d5-8dad959dee42 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.776151] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c9aac1-56de-4d82-b19c-ecfdf9899a92 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.782598] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adba11c-dda4-4109-b8de-8c69db085cef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.795377] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46d7438-4e20-442e-9110-8951eff95cbf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.801195] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eaa5ed9-b96f-4323-ba42-cc87e391a228 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.831170] env[63197]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-9f27a4d2-efb5-40eb-a8d1-28d01357678a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.836306] env[63197]: DEBUG nova.virt.vmwareapi.driver [None req-111d852c-d1ef-4406-9c86-0c7cfa0efbed None None] Extension org.openstack.compute already exists. {{(pid=63197) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 511.839142] env[63197]: INFO nova.compute.provider_config [None req-111d852c-d1ef-4406-9c86-0c7cfa0efbed None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 512.342788] env[63197]: DEBUG nova.context [None req-111d852c-d1ef-4406-9c86-0c7cfa0efbed None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),8eb63ef8-7b9b-4af2-9b3a-09d0c50b822e(cell1) {{(pid=63197) load_cells /opt/stack/nova/nova/context.py:464}} [ 512.344916] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.345151] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.345821] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.346428] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Acquiring lock "8eb63ef8-7b9b-4af2-9b3a-09d0c50b822e" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.346637] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Lock "8eb63ef8-7b9b-4af2-9b3a-09d0c50b822e" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.347683] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Lock "8eb63ef8-7b9b-4af2-9b3a-09d0c50b822e" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.368740] env[63197]: INFO dbcounter [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Registered counter for database nova_cell0 [ 512.376858] env[63197]: INFO dbcounter [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Registered counter for database nova_cell1 [ 512.380301] env[63197]: DEBUG oslo_db.sqlalchemy.engines [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63197) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 512.380906] env[63197]: DEBUG oslo_db.sqlalchemy.engines [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63197) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 512.385407] env[63197]: ERROR nova.db.main.api [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 512.385407] env[63197]: result = function(*args, **kwargs) [ 512.385407] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 512.385407] env[63197]: return func(*args, **kwargs) [ 512.385407] env[63197]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 512.385407] env[63197]: result = fn(*args, **kwargs) [ 512.385407] env[63197]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 512.385407] env[63197]: return f(*args, **kwargs) [ 512.385407] env[63197]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 512.385407] env[63197]: return db.service_get_minimum_version(context, binaries) [ 512.385407] env[63197]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 512.385407] env[63197]: _check_db_access() [ 512.385407] env[63197]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 512.385407] env[63197]: stacktrace = ''.join(traceback.format_stack()) [ 512.385407] env[63197]: [ 512.386437] env[63197]: ERROR nova.db.main.api [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 512.386437] env[63197]: result = function(*args, **kwargs) [ 512.386437] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 512.386437] env[63197]: return func(*args, **kwargs) [ 512.386437] env[63197]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 512.386437] env[63197]: result = fn(*args, **kwargs) [ 512.386437] env[63197]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 512.386437] env[63197]: return f(*args, **kwargs) [ 512.386437] env[63197]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 512.386437] env[63197]: return db.service_get_minimum_version(context, binaries) [ 512.386437] env[63197]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 512.386437] env[63197]: _check_db_access() [ 512.386437] env[63197]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 512.386437] env[63197]: stacktrace = ''.join(traceback.format_stack()) [ 512.386437] env[63197]: [ 512.386909] env[63197]: WARNING nova.objects.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 512.386973] env[63197]: WARNING nova.objects.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Failed to get minimum service version for cell 8eb63ef8-7b9b-4af2-9b3a-09d0c50b822e [ 512.387405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Acquiring lock "singleton_lock" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 512.387569] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Acquired lock "singleton_lock" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 512.387813] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Releasing lock "singleton_lock" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.388181] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Full set of CONF: {{(pid=63197) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 512.388327] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ******************************************************************************** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 512.388454] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Configuration options gathered from: {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 512.388591] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 512.388786] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 512.388925] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ================================================================================ {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 512.389144] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] allow_resize_to_same_host = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.389318] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] arq_binding_timeout = 300 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.389452] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] backdoor_port = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.389581] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] backdoor_socket = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.389746] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] block_device_allocate_retries = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.389905] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] block_device_allocate_retries_interval = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.390090] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cert = self.pem {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.390266] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.390432] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute_monitors = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.390602] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] config_dir = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.390772] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] config_drive_format = iso9660 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.390907] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.391088] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] config_source = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.391261] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] console_host = devstack {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.391426] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] control_exchange = nova {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.391585] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cpu_allocation_ratio = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.391745] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] daemon = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.391910] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] debug = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.392082] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] default_access_ip_network_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.392254] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] default_availability_zone = nova {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.392410] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] default_ephemeral_format = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.392570] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] default_green_pool_size = 1000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.392810] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.392977] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] default_schedule_zone = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.393151] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] disk_allocation_ratio = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.393310] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] enable_new_services = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.393490] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] enabled_apis = ['osapi_compute'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.393650] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] enabled_ssl_apis = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.393807] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] flat_injected = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.393961] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] force_config_drive = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.394136] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] force_raw_images = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.394337] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] graceful_shutdown_timeout = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.394505] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] heal_instance_info_cache_interval = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.394719] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] host = cpu-1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.394896] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.395072] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.395240] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.395450] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.395614] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instance_build_timeout = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.395774] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instance_delete_interval = 300 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.395944] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instance_format = [instance: %(uuid)s] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.396121] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instance_name_template = instance-%08x {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.396283] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instance_usage_audit = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.396450] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instance_usage_audit_period = month {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.396613] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.396777] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.396941] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] internal_service_availability_zone = internal {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.397113] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] key = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.397274] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] live_migration_retry_count = 30 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.397443] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_color = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.397606] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_config_append = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.397770] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.397969] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_dir = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.398122] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.398255] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_options = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.398419] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_rotate_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.398587] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_rotate_interval_type = days {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.398752] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] log_rotation_type = none {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.398882] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.399017] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.399195] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.399363] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.399493] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.399659] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] long_rpc_timeout = 1800 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.399819] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] max_concurrent_builds = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.399979] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] max_concurrent_live_migrations = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.400154] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] max_concurrent_snapshots = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.400315] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] max_local_block_devices = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.400473] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] max_logfile_count = 30 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.400629] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] max_logfile_size_mb = 200 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.400787] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] maximum_instance_delete_attempts = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.400956] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] metadata_listen = 0.0.0.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.401144] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] metadata_listen_port = 8775 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.401313] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] metadata_workers = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.401474] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] migrate_max_retries = -1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.401640] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] mkisofs_cmd = genisoimage {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.401844] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.401979] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] my_ip = 10.180.1.21 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.402158] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] network_allocate_retries = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.402334] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.402501] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.402664] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] osapi_compute_listen_port = 8774 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.402828] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] osapi_compute_unique_server_name_scope = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.402992] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] osapi_compute_workers = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.403171] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] password_length = 12 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.403328] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] periodic_enable = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.403487] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] periodic_fuzzy_delay = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.403654] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] pointer_model = usbtablet {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.403817] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] preallocate_images = none {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.403976] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] publish_errors = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.404120] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] pybasedir = /opt/stack/nova {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.404275] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ram_allocation_ratio = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.404431] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] rate_limit_burst = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.404595] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] rate_limit_except_level = CRITICAL {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.404754] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] rate_limit_interval = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.404912] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] reboot_timeout = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.405083] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] reclaim_instance_interval = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.405245] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] record = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.405412] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] reimage_timeout_per_gb = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.405576] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] report_interval = 120 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.405735] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] rescue_timeout = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.405912] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] reserved_host_cpus = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.406093] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] reserved_host_disk_mb = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.406259] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] reserved_host_memory_mb = 512 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.406420] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] reserved_huge_pages = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.406580] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] resize_confirm_window = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.406740] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] resize_fs_using_block_device = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.406896] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] resume_guests_state_on_host_boot = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.407123] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.407238] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] rpc_response_timeout = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.407399] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] run_external_periodic_tasks = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.407567] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] running_deleted_instance_action = reap {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.407727] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.407885] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] running_deleted_instance_timeout = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.408085] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler_instance_sync_interval = 120 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.408267] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_down_time = 720 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.408437] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] servicegroup_driver = db {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.408593] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] shell_completion = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.408751] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] shelved_offload_time = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.408917] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] shelved_poll_interval = 3600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.409138] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] shutdown_timeout = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.409309] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] source_is_ipv6 = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.409467] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ssl_only = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.409714] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.409882] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] sync_power_state_interval = 600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.410056] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] sync_power_state_pool_size = 1000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.410233] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] syslog_log_facility = LOG_USER {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.410387] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] tempdir = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.410544] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] timeout_nbd = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.410708] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] transport_url = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.410867] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] update_resources_interval = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.411035] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] use_cow_images = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.411199] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] use_eventlog = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.411354] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] use_journal = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.411511] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] use_json = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.411668] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] use_rootwrap_daemon = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.411824] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] use_stderr = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.411981] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] use_syslog = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.412149] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vcpu_pin_set = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.412314] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plugging_is_fatal = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.412480] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plugging_timeout = 300 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.412648] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] virt_mkfs = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.412808] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] volume_usage_poll_interval = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.412968] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] watch_log_file = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.413150] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] web = /usr/share/spice-html5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 512.413331] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.413498] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.413661] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.413828] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_concurrency.disable_process_locking = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.414390] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.414588] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.414761] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.414938] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.415130] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.415302] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.415485] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.auth_strategy = keystone {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.415654] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.compute_link_prefix = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.415842] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.416054] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.dhcp_domain = novalocal {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.416234] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.enable_instance_password = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.416401] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.glance_link_prefix = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.416566] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.416737] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.416900] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.instance_list_per_project_cells = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.417072] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.list_records_by_skipping_down_cells = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.417242] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.local_metadata_per_cell = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.417414] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.max_limit = 1000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.417581] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.metadata_cache_expiration = 15 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.417754] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.neutron_default_tenant_id = default {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.417955] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.response_validation = warn {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.418153] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.use_neutron_default_nets = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.418333] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.418499] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.418669] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.418844] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.419028] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.vendordata_dynamic_targets = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.419204] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.vendordata_jsonfile_path = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.419385] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.419578] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.backend = dogpile.cache.memcached {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.419745] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.backend_argument = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.419922] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.config_prefix = cache.oslo {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.420098] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.dead_timeout = 60.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.420283] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.debug_cache_backend = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.420459] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.enable_retry_client = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.420622] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.enable_socket_keepalive = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.420792] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.enabled = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.420955] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.enforce_fips_mode = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.421139] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.expiration_time = 600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.421302] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.hashclient_retry_attempts = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.421466] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.421626] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_dead_retry = 300 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.421782] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_password = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.421943] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.422118] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.422285] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_pool_maxsize = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.422448] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.422609] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_sasl_enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.422785] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.422950] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.423128] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.memcache_username = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.423295] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.proxies = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.423460] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.redis_db = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.423618] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.redis_password = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.423788] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.423966] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.424153] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.redis_server = localhost:6379 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.424320] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.redis_socket_timeout = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.424479] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.redis_username = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.424642] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.retry_attempts = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.424806] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.retry_delay = 0.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.424970] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.socket_keepalive_count = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.425150] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.socket_keepalive_idle = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.425312] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.socket_keepalive_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.425471] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.tls_allowed_ciphers = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.425628] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.tls_cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.425784] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.tls_certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.425979] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.tls_enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.426161] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cache.tls_keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.426335] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.426510] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.auth_type = password {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.426671] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.426845] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.427015] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.427187] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.427351] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.cross_az_attach = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.427513] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.debug = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.427670] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.endpoint_template = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.427832] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.http_retries = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.428022] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.428197] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.428374] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.os_region_name = RegionOne {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.428539] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.428698] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cinder.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.428869] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.429041] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.cpu_dedicated_set = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.429208] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.cpu_shared_set = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.429377] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.image_type_exclude_list = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.429535] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.429696] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.429862] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.430063] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.430300] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.430481] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.resource_provider_association_refresh = 300 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.430649] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.430842] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.shutdown_retry_interval = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.431096] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.431404] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] conductor.workers = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.431640] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] console.allowed_origins = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.431854] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] console.ssl_ciphers = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.432089] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] console.ssl_minimum_version = default {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.432343] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] consoleauth.enforce_session_timeout = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.432636] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] consoleauth.token_ttl = 600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.432860] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.433159] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.433447] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.433701] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.connect_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.433940] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.connect_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.434206] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.endpoint_override = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.434474] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.434725] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.434964] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.max_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.435216] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.min_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.435435] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.region_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.435613] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.retriable_status_codes = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.435778] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.service_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.436019] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.service_type = accelerator {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.436265] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.436443] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.status_code_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.436605] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.status_code_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.436762] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.436946] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.437160] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] cyborg.version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.437447] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.backend = sqlalchemy {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.437672] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.connection = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.437857] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.connection_debug = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.438082] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.connection_parameters = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.438307] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.connection_recycle_time = 3600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.438572] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.connection_trace = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.438750] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.db_inc_retry_interval = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.438920] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.db_max_retries = 20 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.439101] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.db_max_retry_interval = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.439267] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.db_retry_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.439430] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.max_overflow = 50 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.439591] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.max_pool_size = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.439750] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.max_retries = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.439919] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.440091] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.mysql_wsrep_sync_wait = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.440254] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.pool_timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.440415] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.retry_interval = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.440570] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.slave_connection = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.440729] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.sqlite_synchronous = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.440887] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] database.use_db_reconnect = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.441075] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.backend = sqlalchemy {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.441249] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.connection = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.441412] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.connection_debug = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.441581] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.connection_parameters = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.441743] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.connection_recycle_time = 3600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.441904] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.connection_trace = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.442082] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.db_inc_retry_interval = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.442245] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.db_max_retries = 20 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.442409] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.db_max_retry_interval = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.442572] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.db_retry_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.442735] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.max_overflow = 50 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.442893] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.max_pool_size = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.443069] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.max_retries = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.443245] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.443404] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.443563] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.pool_timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.443723] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.retry_interval = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.443882] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.slave_connection = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.444055] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] api_database.sqlite_synchronous = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.444232] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] devices.enabled_mdev_types = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.444444] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.444668] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.444843] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ephemeral_storage_encryption.enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.445020] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.445193] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.api_servers = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.445361] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.445522] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.445686] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.445847] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.connect_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.446014] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.connect_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.446187] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.debug = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.446354] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.default_trusted_certificate_ids = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.446517] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.enable_certificate_validation = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.446680] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.enable_rbd_download = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.446838] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.endpoint_override = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.447008] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.447193] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.447352] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.max_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.447557] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.min_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.447670] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.num_retries = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.447837] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.rbd_ceph_conf = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.448018] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.rbd_connect_timeout = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.448199] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.rbd_pool = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.448365] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.rbd_user = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.448526] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.region_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.448684] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.retriable_status_codes = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.448839] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.service_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.449025] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.service_type = image {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.449183] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.449346] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.status_code_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.449501] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.status_code_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.449652] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.449829] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.449989] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.verify_glance_signatures = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.450163] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] glance.version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.450330] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] guestfs.debug = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.450491] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] mks.enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.450846] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.451050] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] image_cache.manager_interval = 2400 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.451228] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] image_cache.precache_concurrency = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.451399] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] image_cache.remove_unused_base_images = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.451573] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.451742] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.451918] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] image_cache.subdirectory_name = _base {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.452108] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.api_max_retries = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.452277] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.api_retry_interval = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.452437] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.452598] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.auth_type = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.452756] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.452912] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.453088] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.453256] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.conductor_group = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.453413] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.connect_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.453570] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.connect_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.453728] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.endpoint_override = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.453887] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.454057] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.454222] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.max_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.454379] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.min_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.454545] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.peer_list = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.454703] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.region_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.454861] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.retriable_status_codes = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.455034] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.serial_console_state_timeout = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.455201] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.service_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.455368] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.service_type = baremetal {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.455527] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.shard = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.455687] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.455842] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.status_code_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.455998] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.status_code_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.456171] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.456350] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.456510] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ironic.version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.456689] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.456861] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] key_manager.fixed_key = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.457059] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.457229] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.barbican_api_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.457385] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.barbican_endpoint = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.457556] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.barbican_endpoint_type = public {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.457717] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.barbican_region_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.457873] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.458055] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.458225] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.458385] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.458547] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.458709] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.number_of_retries = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.458869] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.retry_delay = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.459048] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.send_service_user_token = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.459220] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.459379] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.459540] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.verify_ssl = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.459697] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican.verify_ssl_path = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.459863] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.460039] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.auth_type = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.460206] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.460366] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.460523] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.460681] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.460838] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.460998] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.461171] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] barbican_service_user.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.461337] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.approle_role_id = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.461494] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.approle_secret_id = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.461660] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.kv_mountpoint = secret {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.461817] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.kv_path = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.461977] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.kv_version = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.462151] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.namespace = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.462308] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.root_token_id = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.462465] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.ssl_ca_crt_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.462628] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.timeout = 60.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.462787] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.use_ssl = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.462955] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.463137] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.463299] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.auth_type = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.463455] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.463608] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.463765] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.463919] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.connect_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.464085] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.connect_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.464246] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.endpoint_override = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.464406] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.464562] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.464715] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.max_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.464868] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.min_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.465032] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.region_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.465194] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.retriable_status_codes = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.465353] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.service_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.465521] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.service_type = identity {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.465682] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.465840] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.status_code_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.465997] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.status_code_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.466171] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.466352] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.466512] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] keystone.version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.466714] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.connection_uri = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.466878] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.cpu_mode = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.467058] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.467237] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.cpu_models = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.467409] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.cpu_power_governor_high = performance {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.467577] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.467772] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.cpu_power_management = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.467912] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.468110] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.device_detach_attempts = 8 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.468285] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.device_detach_timeout = 20 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.468453] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.disk_cachemodes = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.468615] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.disk_prefix = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.468787] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.enabled_perf_events = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.468965] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.file_backed_memory = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.469173] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.gid_maps = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.469343] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.hw_disk_discard = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.469503] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.hw_machine_type = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.469678] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.images_rbd_ceph_conf = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.469847] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.470019] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.470197] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.images_rbd_glance_store_name = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.470372] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.images_rbd_pool = rbd {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.470540] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.images_type = default {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.470697] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.images_volume_group = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.470857] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.inject_key = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.471027] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.inject_partition = -2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.471199] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.inject_password = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.471356] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.iscsi_iface = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.471515] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.iser_use_multipath = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.471676] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.471841] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.472013] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_downtime = 500 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.472184] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.472346] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.472511] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_inbound_addr = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.472673] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.472838] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.473008] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_scheme = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.473196] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_timeout_action = abort {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.473389] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_tunnelled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.473559] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_uri = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.473722] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.live_migration_with_native_tls = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.473883] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.max_queues = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.474060] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.474303] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.474468] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.nfs_mount_options = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.474767] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.474944] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.475127] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.475292] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.475456] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.475619] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.num_pcie_ports = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.475788] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.475955] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.pmem_namespaces = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.476131] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.quobyte_client_cfg = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.476493] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.476617] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.476786] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.476951] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.477132] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rbd_secret_uuid = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.477296] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rbd_user = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.477460] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.477632] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.477794] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rescue_image_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.477977] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rescue_kernel_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.478162] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rescue_ramdisk_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.478337] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.478499] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.rx_queue_size = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.478669] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.smbfs_mount_options = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.478950] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.479277] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.snapshot_compression = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.479437] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.snapshot_image_format = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.479669] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.479839] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.sparse_logical_volumes = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.480017] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.swtpm_enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.480201] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.swtpm_group = tss {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.480374] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.swtpm_user = tss {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.480572] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.sysinfo_serial = unique {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.480697] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.tb_cache_size = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.480859] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.tx_queue_size = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.481035] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.uid_maps = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.481244] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.use_virtio_for_bridges = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.481507] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.virt_type = kvm {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.481717] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.volume_clear = zero {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.481890] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.volume_clear_size = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.482074] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.volume_use_multipath = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.482246] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.vzstorage_cache_path = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.482419] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.482590] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.482758] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.482955] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.483256] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.483437] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.vzstorage_mount_user = stack {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.483603] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.483778] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.483952] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.auth_type = password {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.484134] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.484296] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.484463] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.484624] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.connect_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.484783] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.connect_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.484954] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.default_floating_pool = public {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.485132] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.endpoint_override = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.485301] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.extension_sync_interval = 600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.485462] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.http_retries = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.485622] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.485778] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.485935] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.max_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.486119] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.486279] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.min_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.486445] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.ovs_bridge = br-int {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.486609] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.physnets = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.486779] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.region_name = RegionOne {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.486938] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.retriable_status_codes = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.487120] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.service_metadata_proxy = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.487283] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.service_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.487450] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.service_type = network {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.487612] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.487767] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.status_code_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.487929] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.status_code_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.488117] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.488306] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.488467] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] neutron.version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.488641] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] notifications.bdms_in_notifications = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.488817] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] notifications.default_level = INFO {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.488995] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] notifications.notification_format = unversioned {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.489202] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] notifications.notify_on_state_change = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.489386] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.489565] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] pci.alias = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.489736] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] pci.device_spec = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.489900] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] pci.report_in_placement = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.490088] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.490269] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.auth_type = password {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.490438] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.490600] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.490760] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.490924] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.491096] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.connect_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.491258] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.connect_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.491415] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.default_domain_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.491572] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.default_domain_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.491729] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.domain_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.491883] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.domain_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.492050] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.endpoint_override = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.492218] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.492374] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.492529] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.max_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.492682] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.min_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.492846] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.password = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.493017] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.project_domain_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.493190] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.project_domain_name = Default {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.493359] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.project_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.493531] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.project_name = service {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.493701] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.region_name = RegionOne {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.493873] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.retriable_status_codes = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.494046] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.service_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.494225] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.service_type = placement {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.494388] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.494549] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.status_code_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.494712] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.status_code_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.494870] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.system_scope = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.495039] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.495205] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.trust_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.495363] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.user_domain_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.495531] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.user_domain_name = Default {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.495690] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.user_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.495864] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.username = nova {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.496056] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.496231] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] placement.version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.496402] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.cores = 20 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.496570] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.count_usage_from_placement = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.496744] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.496920] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.injected_file_content_bytes = 10240 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.497107] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.injected_file_path_length = 255 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.497277] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.injected_files = 5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.497446] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.instances = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.497611] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.key_pairs = 100 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.497776] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.metadata_items = 128 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.497973] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.ram = 51200 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.498177] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.recheck_quota = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.498335] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.server_group_members = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.498500] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] quota.server_groups = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.498674] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.498838] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.499040] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.image_metadata_prefilter = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.499210] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.499379] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.max_attempts = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.499543] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.max_placement_results = 1000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.499707] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.499870] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.500064] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.500219] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] scheduler.workers = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.500395] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.500568] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.500766] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.500921] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.501099] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.501268] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.501430] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.501623] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.501793] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.host_subset_size = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.501963] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.502136] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.502300] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.502466] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.isolated_hosts = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.502632] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.isolated_images = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.502797] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.502959] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.503137] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.503303] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.pci_in_placement = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.503465] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.503628] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.503791] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.503953] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.504131] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.504298] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.504462] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.track_instance_changes = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.504639] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.504811] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] metrics.required = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.504977] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] metrics.weight_multiplier = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.505161] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.505327] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] metrics.weight_setting = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.505647] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.505841] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] serial_console.enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.506013] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] serial_console.port_range = 10000:20000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.506194] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.506361] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.506531] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] serial_console.serialproxy_port = 6083 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.506701] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.506875] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.auth_type = password {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.507047] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.507216] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.507380] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.507542] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.507700] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.507870] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.send_service_user_token = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.508069] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.508297] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] service_user.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.508426] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.agent_enabled = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.508589] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.508897] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.509126] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.509319] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.html5proxy_port = 6082 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.509487] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.image_compression = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.509650] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.jpeg_compression = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.509812] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.playback_compression = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.509976] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.require_secure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.510167] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.server_listen = 127.0.0.1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.510346] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.510498] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.streaming_mode = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.510656] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] spice.zlib_compression = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.510866] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] upgrade_levels.baseapi = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.510992] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] upgrade_levels.compute = auto {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.511167] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] upgrade_levels.conductor = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.511326] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] upgrade_levels.scheduler = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.511490] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.511651] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.511807] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.511963] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.512141] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.512301] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.512458] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.512619] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.512774] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vendordata_dynamic_auth.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.512949] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.api_retry_count = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.513124] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.ca_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.513297] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.513461] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.cluster_name = testcl1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.513623] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.connection_pool_size = 10 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.513780] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.console_delay_seconds = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.513950] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.datastore_regex = ^datastore.* {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.514175] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.514350] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.host_password = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.514519] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.host_port = 443 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.514687] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.host_username = administrator@vsphere.local {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.514853] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.insecure = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.515136] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.integration_bridge = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.515209] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.maximum_objects = 100 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.515355] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.pbm_default_policy = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.515514] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.pbm_enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.515669] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.pbm_wsdl_location = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.515836] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.515993] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.serial_port_proxy_uri = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.516169] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.serial_port_service_uri = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.516335] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.task_poll_interval = 0.5 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.516503] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.use_linked_clone = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.516668] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.vnc_keymap = en-us {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.516832] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.vnc_port = 5900 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.516994] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vmware.vnc_port_total = 10000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.517193] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.auth_schemes = ['none'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.517365] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.517655] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.517838] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.518044] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.novncproxy_port = 6080 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.518236] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.server_listen = 127.0.0.1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.518411] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.518572] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.vencrypt_ca_certs = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.518728] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.vencrypt_client_cert = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.518894] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vnc.vencrypt_client_key = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.519094] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.519278] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.disable_deep_image_inspection = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.519440] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.519600] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.519761] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.519919] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.disable_rootwrap = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.520095] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.enable_numa_live_migration = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.520259] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.520421] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.520582] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.520761] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.libvirt_disable_apic = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.520966] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.521111] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.521268] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.521430] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.521591] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.521749] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.521906] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.522076] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.522240] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.522404] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.522586] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.522754] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.client_socket_timeout = 900 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.522920] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.default_pool_size = 1000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.523099] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.keep_alive = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.523267] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.max_header_line = 16384 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.523428] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.523587] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.ssl_ca_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.523746] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.ssl_cert_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.523904] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.ssl_key_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.524080] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.tcp_keepidle = 600 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.524273] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.524441] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] zvm.ca_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.524601] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] zvm.cloud_connector_url = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.524891] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.525078] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] zvm.reachable_timeout = 300 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.525261] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.enforce_new_defaults = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.525656] env[63197]: WARNING oslo_config.cfg [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 512.525842] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.enforce_scope = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.526023] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.policy_default_rule = default {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.526212] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.526392] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.policy_file = policy.yaml {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.526564] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.526727] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.526888] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.527058] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.527228] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.527396] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.527572] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.527748] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.connection_string = messaging:// {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.527926] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.enabled = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.528127] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.es_doc_type = notification {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.528304] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.es_scroll_size = 10000 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.528527] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.es_scroll_time = 2m {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.528636] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.filter_error_trace = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.528804] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.hmac_keys = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.528970] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.sentinel_service_name = mymaster {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.529176] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.socket_timeout = 0.1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.529346] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.trace_requests = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.529507] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler.trace_sqlalchemy = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.529691] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler_jaeger.process_tags = {} {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.529851] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler_jaeger.service_name_prefix = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.530023] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] profiler_otlp.service_name_prefix = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.530193] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] remote_debug.host = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.530354] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] remote_debug.port = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.530533] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.530696] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.530859] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.531074] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.531194] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.531361] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.531523] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.531684] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.531844] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.532017] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.532181] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.532355] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.532522] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.532692] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.532862] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.533040] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.533210] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.533382] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.533545] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.533705] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.533870] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.534045] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.534216] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.534375] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.534535] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.534698] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.534860] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.535031] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.535205] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.535371] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.ssl = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.535543] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.535711] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.535876] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.536056] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.536233] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.536395] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.536581] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.536747] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_notifications.retry = -1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.536930] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.537123] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.537296] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.auth_section = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.537459] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.auth_type = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.537620] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.cafile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.537777] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.certfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.537955] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.collect_timing = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.538151] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.connect_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.538316] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.connect_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.538477] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.endpoint_id = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.538643] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.endpoint_override = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.538796] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.insecure = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.538955] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.keyfile = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.539155] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.max_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.539327] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.min_version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.539488] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.region_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.539710] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.retriable_status_codes = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.539896] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.service_name = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.540072] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.service_type = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.540242] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.split_loggers = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.540401] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.status_code_retries = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.540561] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.status_code_retry_delay = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.540719] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.timeout = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.540879] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.valid_interfaces = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.541049] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_limit.version = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.541223] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_reports.file_event_handler = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.541389] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.541548] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] oslo_reports.log_dir = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.541718] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.541878] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.542049] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.542223] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.542390] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.542550] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.542718] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.543022] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_ovs_privileged.group = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.543202] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.543375] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.543605] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.543788] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] vif_plug_ovs_privileged.user = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.543966] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.544170] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.544349] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.544519] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.544690] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.544856] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.545037] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.545208] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.545391] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.545586] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_ovs.isolate_vif = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.545836] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.546102] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.546304] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.546481] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.546648] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] os_vif_ovs.per_port_bridge = False {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.546820] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] privsep_osbrick.capabilities = [21] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.546982] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] privsep_osbrick.group = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.547159] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] privsep_osbrick.helper_command = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.547326] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.547490] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.547648] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] privsep_osbrick.user = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.547823] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.548018] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] nova_sys_admin.group = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.548199] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] nova_sys_admin.helper_command = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.548369] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.548532] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.548757] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] nova_sys_admin.user = None {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 512.548822] env[63197]: DEBUG oslo_service.service [None req-3131bf27-d7b0-4581-be2f-f76116c571e4 None None] ******************************************************************************** {{(pid=63197) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 512.549379] env[63197]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 513.052514] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Getting list of instances from cluster (obj){ [ 513.052514] env[63197]: value = "domain-c8" [ 513.052514] env[63197]: _type = "ClusterComputeResource" [ 513.052514] env[63197]: } {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 513.053712] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d16883b-4b8a-4e60-b316-8a8996831824 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.062491] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Got total of 0 instances {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 513.063014] env[63197]: WARNING nova.virt.vmwareapi.driver [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 513.063480] env[63197]: INFO nova.virt.node [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Generated node identity e9e01824-bb93-4102-9e5b-66d5dbc1990b [ 513.063708] env[63197]: INFO nova.virt.node [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Wrote node identity e9e01824-bb93-4102-9e5b-66d5dbc1990b to /opt/stack/data/n-cpu-1/compute_id [ 513.566848] env[63197]: WARNING nova.compute.manager [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Compute nodes ['e9e01824-bb93-4102-9e5b-66d5dbc1990b'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 514.616562] env[63197]: INFO nova.compute.manager [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 515.623110] env[63197]: WARNING nova.compute.manager [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 515.623494] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.623591] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.623741] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.623895] env[63197]: DEBUG nova.compute.resource_tracker [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 515.624835] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e398ec-9a83-4053-aaff-ba49fa9ae595 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.633162] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548344b4-73a3-4f46-86ee-2666896aef26 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.646071] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0f2432-1da3-4514-a824-3e3bc98847e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.652491] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1d2bd6-0277-46b6-9e05-30d8260fb7e1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.684574] env[63197]: DEBUG nova.compute.resource_tracker [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181501MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 515.684705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.684911] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.187841] env[63197]: WARNING nova.compute.resource_tracker [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] No compute node record for cpu-1:e9e01824-bb93-4102-9e5b-66d5dbc1990b: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host e9e01824-bb93-4102-9e5b-66d5dbc1990b could not be found. [ 516.691655] env[63197]: INFO nova.compute.resource_tracker [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: e9e01824-bb93-4102-9e5b-66d5dbc1990b [ 518.200759] env[63197]: DEBUG nova.compute.resource_tracker [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 518.201097] env[63197]: DEBUG nova.compute.resource_tracker [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 518.349544] env[63197]: INFO nova.scheduler.client.report [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] [req-83827360-7919-4bff-b0e5-9dda20931dd6] Created resource provider record via placement API for resource provider with UUID e9e01824-bb93-4102-9e5b-66d5dbc1990b and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 518.367476] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c07a8fa-4c37-4281-b656-30ba16c73f4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.375117] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c16e75-b1b1-4607-9710-f87e0dbfca7b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.404622] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a87ab3-dbff-4f52-a52f-cbc53c2419d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.411524] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6eef85-b4c4-400f-a96a-e9c2bc2fef6b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.424330] env[63197]: DEBUG nova.compute.provider_tree [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 518.963135] env[63197]: DEBUG nova.scheduler.client.report [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Updated inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 518.963372] env[63197]: DEBUG nova.compute.provider_tree [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Updating resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b generation from 0 to 1 during operation: update_inventory {{(pid=63197) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 518.963512] env[63197]: DEBUG nova.compute.provider_tree [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 519.011877] env[63197]: DEBUG nova.compute.provider_tree [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Updating resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b generation from 1 to 2 during operation: update_traits {{(pid=63197) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 519.516584] env[63197]: DEBUG nova.compute.resource_tracker [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 519.516944] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.832s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.516944] env[63197]: DEBUG nova.service [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Creating RPC server for service compute {{(pid=63197) start /opt/stack/nova/nova/service.py:186}} [ 519.530539] env[63197]: DEBUG nova.service [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] Join ServiceGroup membership for this service compute {{(pid=63197) start /opt/stack/nova/nova/service.py:203}} [ 519.530730] env[63197]: DEBUG nova.servicegroup.drivers.db [None req-cf6f155b-b32d-47bd-a276-58311c83583c None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63197) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 557.929891] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "ac9d308b-ac4b-4153-8e1d-7551213c8233" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.930179] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "ac9d308b-ac4b-4153-8e1d-7551213c8233" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.435809] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 558.975098] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.975098] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.976052] env[63197]: INFO nova.compute.claims [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 559.490893] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.490893] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.935709] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Acquiring lock "7ec7941d-183d-41d6-b2fb-6e795786c454" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.935709] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Lock "7ec7941d-183d-41d6-b2fb-6e795786c454" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.005531] env[63197]: DEBUG nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.020530] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Acquiring lock "a892c650-d8e8-4eaa-b025-cf330478aabf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.020530] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Lock "a892c650-d8e8-4eaa-b025-cf330478aabf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.108570] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4c2541-d3cc-4fc5-bd1f-055035857fdf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.120174] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9f7ebb-39d8-4108-9b88-7818af5e829c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.156718] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b511c5ef-ac65-4d08-93ed-8c3a220b13ea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.165340] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440f568b-8d1e-4564-8e8f-7427126ab9c8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.181676] env[63197]: DEBUG nova.compute.provider_tree [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.439489] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.523405] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.534164] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.685856] env[63197]: DEBUG nova.scheduler.client.report [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.814512] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.815045] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.971557] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.057547] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.191162] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.217s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.193632] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 561.196947] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.663s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.200301] env[63197]: INFO nova.compute.claims [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.317768] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 561.405524] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Acquiring lock "c0742ebb-c176-44cf-ae13-d778cc6364cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.406886] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Lock "c0742ebb-c176-44cf-ae13-d778cc6364cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.699807] env[63197]: DEBUG nova.compute.utils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.701278] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 561.701560] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 561.852196] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.909453] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.137545] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "99539d7f-cd93-4250-97cd-2449e03e517d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.137900] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "99539d7f-cd93-4250-97cd-2449e03e517d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.206392] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 562.223770] env[63197]: DEBUG nova.policy [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94419f4b0e6346ea830b0072ab80f798', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '969bf888642249289ef4a81d5492d698', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 562.387284] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d8fc11-2c10-43a8-8b9a-cbca483c057c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.399096] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637e04ba-2528-4183-88a9-f2bf0277ab7a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.437883] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fa0ae9-b2e9-4c3c-96e6-1fc798a0f2ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.452345] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01cf687-2a16-455f-9e48-11f536fd2123 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.466440] env[63197]: DEBUG nova.compute.provider_tree [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.468673] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.484273] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Acquiring lock "68ce48d7-e967-4cfc-a607-f6580bc14648" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.484273] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Lock "68ce48d7-e967-4cfc-a607-f6580bc14648" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.640213] env[63197]: DEBUG nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.970718] env[63197]: DEBUG nova.scheduler.client.report [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.991772] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.170080] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Successfully created port: a51c1f11-407a-4fd4-be72-5a92a7d3a456 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.177410] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.240931] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 563.276329] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.276443] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.276505] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.276685] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.277523] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.277523] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.277523] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.277523] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.277784] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.277898] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.278209] env[63197]: DEBUG nova.virt.hardware [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.280394] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20c02f8-cd14-48e1-a3e3-d6114e7e7731 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.291271] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdc41c4-31f8-4a69-bfb2-98b693f06758 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.308489] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879438e2-5c9b-4f14-883e-9dbfe535c113 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.479770] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.480434] env[63197]: DEBUG nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.484994] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.514s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.486714] env[63197]: INFO nova.compute.claims [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 563.527960] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.990974] env[63197]: DEBUG nova.compute.utils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.996779] env[63197]: DEBUG nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Not allocating networking since 'none' was specified. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 564.501052] env[63197]: DEBUG nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 564.677439] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8aa430-c5d3-4e17-92c6-fd618a8b95d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.689118] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98406eca-199f-4085-8200-f64d408d2c2e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.725671] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac78ba8-367f-47da-84e3-8ec386957ef7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.734444] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6998c5-cf0c-4f75-ade9-14a0d560bca2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.749619] env[63197]: DEBUG nova.compute.provider_tree [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.253018] env[63197]: DEBUG nova.scheduler.client.report [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.520738] env[63197]: DEBUG nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.572542] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.572929] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.573176] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.574788] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.575085] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.575263] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.575486] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.575647] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.575811] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.575966] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.576144] env[63197]: DEBUG nova.virt.hardware [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.577544] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a498947-8163-4cc5-abf0-3324dbb20c7d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.587007] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6ce9ff-ce2b-4108-9cb0-11321637cd5c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.601288] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 565.612805] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 565.612935] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01bbff68-464b-4caa-b542-fd9bf09cdf2e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.627731] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Created folder: OpenStack in parent group-v4. [ 565.627980] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Creating folder: Project (911f0fd814d34b108c9c219bec62def4). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 565.628258] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f01019a5-7c4a-4f50-8aa6-c2101c3b531e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.639734] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Created folder: Project (911f0fd814d34b108c9c219bec62def4) in parent group-v290286. [ 565.640142] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Creating folder: Instances. Parent ref: group-v290287. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 565.641743] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05a28768-1bf3-481a-9f5e-2a46e8fb21f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.651199] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Created folder: Instances in parent group-v290287. [ 565.651696] env[63197]: DEBUG oslo.service.loopingcall [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.651904] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 565.652222] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf3c7095-9090-4751-912e-2a9d7b1fe4f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.671045] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 565.671045] env[63197]: value = "task-1364006" [ 565.671045] env[63197]: _type = "Task" [ 565.671045] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.680574] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364006, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.762020] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.275s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.762020] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 565.771113] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.712s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.772343] env[63197]: INFO nova.compute.claims [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.805751] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Acquiring lock "37280bd4-3cd8-4910-97de-0cc7494abc22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.806055] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Lock "37280bd4-3cd8-4910-97de-0cc7494abc22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.182980] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364006, 'name': CreateVM_Task, 'duration_secs': 0.30392} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.183341] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 566.184480] env[63197]: DEBUG oslo_vmware.service [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15205fe8-6614-4a0d-a3f7-df000eb78261 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.190988] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.191520] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.192389] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 566.192704] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72fdcb94-d1e9-47c0-a2b7-2389308e589d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.198054] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 566.198054] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528e703c-956c-a7b5-7330-c8ef1435ac7d" [ 566.198054] env[63197]: _type = "Task" [ 566.198054] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.207904] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528e703c-956c-a7b5-7330-c8ef1435ac7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.278933] env[63197]: DEBUG nova.compute.utils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 566.280329] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 566.280562] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 566.314850] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 566.431752] env[63197]: DEBUG nova.policy [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '470cbd32e6284fc09663aee33e78cfb9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ee53af2d03f4257827fa634c80e3e61', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 566.715233] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.715233] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 566.715233] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.715233] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.715776] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 566.715776] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4d77280e-9de3-4261-a665-2dd195393db5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.734962] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 566.734962] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 566.734962] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a5df18-58e4-47f2-872d-5dcf87dab9b0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.743573] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1133e60f-9ff2-40ab-b6c1-3545a364f442 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.751100] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 566.751100] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527b92cc-76ea-0744-b1f7-1f8aadc58e0a" [ 566.751100] env[63197]: _type = "Task" [ 566.751100] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.763397] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527b92cc-76ea-0744-b1f7-1f8aadc58e0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.792467] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 566.850247] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.997113] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c244b040-2bde-47be-9bbd-c1268db5edbe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.011892] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74179ca1-729c-40b9-b766-e32b6865b03f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.053589] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e575fc64-3185-4f50-b062-06793d0f7d53 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.061706] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a59870-bfe1-4782-85e6-8cf00289e5a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.080200] env[63197]: DEBUG nova.compute.provider_tree [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.264828] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Preparing fetch location {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 567.264828] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Creating directory with path [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 567.264828] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7de2fb7c-f01a-4ed4-91ba-78955a886159 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.293120] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Created directory with path [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 567.293120] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Fetch image to [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 567.293120] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Downloading image file data f760fedc-0b5b-4c56-acbe-239b47c945e8 to [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk on the data store datastore2 {{(pid=63197) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 567.293120] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7daa19-bf0e-4d4d-a063-2cac7a7ff25e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.310047] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eba6c37-bedd-4d55-99d0-8fef6a043367 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.324435] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d5746b-8fb4-4fb1-b40c-f1389816c2a6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.363792] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c97a2f55-62b5-4e51-b0ae-ffbe344162aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.373059] env[63197]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-066d66d3-88d6-4689-b560-a9520ad62cf2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.407218] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Downloading image file data f760fedc-0b5b-4c56-acbe-239b47c945e8 to the data store datastore2 {{(pid=63197) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 567.505586] env[63197]: DEBUG oslo_vmware.rw_handles [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63197) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 567.585926] env[63197]: DEBUG nova.scheduler.client.report [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.590767] env[63197]: ERROR nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. [ 567.590767] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 567.590767] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.590767] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 567.590767] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.590767] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 567.590767] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.590767] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 567.590767] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.590767] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 567.590767] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.590767] env[63197]: ERROR nova.compute.manager raise self.value [ 567.590767] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.590767] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 567.590767] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.590767] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 567.591322] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.591322] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 567.591322] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. [ 567.591322] env[63197]: ERROR nova.compute.manager [ 567.591322] env[63197]: Traceback (most recent call last): [ 567.591322] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 567.591322] env[63197]: listener.cb(fileno) [ 567.591322] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.591322] env[63197]: result = function(*args, **kwargs) [ 567.591322] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 567.591322] env[63197]: return func(*args, **kwargs) [ 567.591322] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.591322] env[63197]: raise e [ 567.591322] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.591322] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 567.591322] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.591322] env[63197]: created_port_ids = self._update_ports_for_instance( [ 567.591322] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.591322] env[63197]: with excutils.save_and_reraise_exception(): [ 567.591322] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.591322] env[63197]: self.force_reraise() [ 567.591322] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.591322] env[63197]: raise self.value [ 567.591322] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.591322] env[63197]: updated_port = self._update_port( [ 567.591322] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.591322] env[63197]: _ensure_no_port_binding_failure(port) [ 567.591322] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.591322] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 567.592197] env[63197]: nova.exception.PortBindingFailed: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. [ 567.592197] env[63197]: Removing descriptor: 15 [ 567.593540] env[63197]: ERROR nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Traceback (most recent call last): [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] yield resources [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self.driver.spawn(context, instance, image_meta, [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] vm_ref = self.build_virtual_machine(instance, [ 567.593540] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] for vif in network_info: [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return self._sync_wrapper(fn, *args, **kwargs) [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self.wait() [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self[:] = self._gt.wait() [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return self._exit_event.wait() [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.593920] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] result = hub.switch() [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return self.greenlet.switch() [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] result = function(*args, **kwargs) [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return func(*args, **kwargs) [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] raise e [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] nwinfo = self.network_api.allocate_for_instance( [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] created_port_ids = self._update_ports_for_instance( [ 567.594271] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] with excutils.save_and_reraise_exception(): [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self.force_reraise() [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] raise self.value [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] updated_port = self._update_port( [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] _ensure_no_port_binding_failure(port) [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] raise exception.PortBindingFailed(port_id=port['id']) [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] nova.exception.PortBindingFailed: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. [ 567.594612] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] [ 567.594969] env[63197]: INFO nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Terminating instance [ 567.599568] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.599726] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquired lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.599936] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 567.605316] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Successfully created port: 8569eb9e-ec1e-4cda-8885-daf63a6ef32e {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 567.815405] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 567.856685] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.857408] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.857553] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.857942] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.857942] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.858128] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.860025] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.860025] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.860025] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.860025] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.860025] env[63197]: DEBUG nova.virt.hardware [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.860270] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f79bb4-b682-4223-8555-11610a8759ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.871802] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff997116-8dc9-4b2d-adb8-8aced45dda09 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.102639] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.102639] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 568.111022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.259s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.113161] env[63197]: INFO nova.compute.claims [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.195917] env[63197]: DEBUG oslo_vmware.rw_handles [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Completed reading data from the image iterator. {{(pid=63197) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 568.197096] env[63197]: DEBUG oslo_vmware.rw_handles [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 568.201489] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.323226] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.337775] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Downloaded image file data f760fedc-0b5b-4c56-acbe-239b47c945e8 to vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk on the data store datastore2 {{(pid=63197) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 568.339728] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Caching image {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 568.339836] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Copying Virtual Disk [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk to [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 568.341550] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01e9b8cf-6847-431c-9fb9-2284a12b0e84 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.354646] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 568.354646] env[63197]: value = "task-1364007" [ 568.354646] env[63197]: _type = "Task" [ 568.354646] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.369351] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.622561] env[63197]: DEBUG nova.compute.utils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.630497] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 568.630608] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.794931] env[63197]: DEBUG nova.policy [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bda8f63db56410988a29aea012f8e50', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f68a2504f37447dfa2ef101da52f3b72', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.826251] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Releasing lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.826568] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 568.826568] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 568.827210] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5514d1ed-5892-40e1-9f66-22687a3bb4c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.838236] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18ba5f5-783e-4825-ac64-1fa05c6650b0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.869181] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ac9d308b-ac4b-4153-8e1d-7551213c8233 could not be found. [ 568.869431] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 568.869860] env[63197]: INFO nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Took 0.04 seconds to destroy the instance on the hypervisor. [ 568.870142] env[63197]: DEBUG oslo.service.loopingcall [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.871483] env[63197]: DEBUG nova.compute.manager [-] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.871580] env[63197]: DEBUG nova.network.neutron [-] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.887634] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364007, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.930162] env[63197]: DEBUG nova.network.neutron [-] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.084065] env[63197]: DEBUG nova.compute.manager [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Received event network-changed-a51c1f11-407a-4fd4-be72-5a92a7d3a456 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.084391] env[63197]: DEBUG nova.compute.manager [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Refreshing instance network info cache due to event network-changed-a51c1f11-407a-4fd4-be72-5a92a7d3a456. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.084663] env[63197]: DEBUG oslo_concurrency.lockutils [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] Acquiring lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.084810] env[63197]: DEBUG oslo_concurrency.lockutils [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] Acquired lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.084974] env[63197]: DEBUG nova.network.neutron [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Refreshing network info cache for port a51c1f11-407a-4fd4-be72-5a92a7d3a456 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 569.131780] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 569.181186] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "a8315374-abb6-4663-bdb5-c4d88cdbbaa4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.181186] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "a8315374-abb6-4663-bdb5-c4d88cdbbaa4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.370716] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364007, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.433834] env[63197]: DEBUG nova.network.neutron [-] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.544018] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02c1f31-4314-4e1b-a4b2-5a1ff133843d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.554946] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1c9c4d-0624-4688-ad6b-c96ef7460a67 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.594841] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f050e987-0cf0-4919-87c1-12356bdef5d6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.606899] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af77a5c9-192e-45bb-b03e-de3bbab0f980 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.625992] env[63197]: DEBUG nova.compute.provider_tree [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.656299] env[63197]: DEBUG nova.network.neutron [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.682382] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 569.869773] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364007, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.937406] env[63197]: INFO nova.compute.manager [-] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Took 1.06 seconds to deallocate network for instance. [ 569.941923] env[63197]: DEBUG nova.compute.claims [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 569.942320] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.042127] env[63197]: DEBUG nova.network.neutron [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.133568] env[63197]: DEBUG nova.scheduler.client.report [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.152181] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 570.212197] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 570.212624] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 570.212624] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 570.212812] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 570.212947] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 570.213104] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 570.213309] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 570.213536] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 570.213632] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 570.213767] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 570.214090] env[63197]: DEBUG nova.virt.hardware [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 570.216597] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24108b18-02a7-4ffa-8fdc-6974f098236d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.225906] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.229413] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d577ff7d-a204-45ef-af4b-531eef801b54 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.322234] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Successfully created port: 9d17dab0-f776-4eb7-8729-0ede74a50ddd {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 570.368357] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364007, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.67124} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.368602] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Copied Virtual Disk [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk to [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 570.368771] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Deleting the datastore file [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 570.369615] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e7911b9-2492-4f2e-9564-706e2ab4b17b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.377681] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 570.377681] env[63197]: value = "task-1364008" [ 570.377681] env[63197]: _type = "Task" [ 570.377681] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.391031] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364008, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.547678] env[63197]: DEBUG oslo_concurrency.lockutils [req-595babc3-f0ad-43bf-ae53-d8c07e809713 req-04fc72bc-42ed-4835-8dac-2805ed6fcc61 service nova] Releasing lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.642228] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.642447] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.645660] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.177s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.647652] env[63197]: INFO nova.compute.claims [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.891307] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364008, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023452} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.894937] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 570.894937] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Moving file from [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd/f760fedc-0b5b-4c56-acbe-239b47c945e8 to [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8. {{(pid=63197) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 570.894937] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a8cb4fa0-3da8-4cdf-a6a0-864661b8a6e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.899769] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 570.899769] env[63197]: value = "task-1364009" [ 570.899769] env[63197]: _type = "Task" [ 570.899769] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.908915] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364009, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.107328] env[63197]: ERROR nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. [ 571.107328] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 571.107328] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.107328] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 571.107328] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.107328] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 571.107328] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.107328] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 571.107328] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.107328] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 571.107328] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.107328] env[63197]: ERROR nova.compute.manager raise self.value [ 571.107328] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.107328] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 571.107328] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.107328] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 571.110572] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.110572] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 571.110572] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. [ 571.110572] env[63197]: ERROR nova.compute.manager [ 571.110572] env[63197]: Traceback (most recent call last): [ 571.110572] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 571.110572] env[63197]: listener.cb(fileno) [ 571.110572] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.110572] env[63197]: result = function(*args, **kwargs) [ 571.110572] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.110572] env[63197]: return func(*args, **kwargs) [ 571.110572] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.110572] env[63197]: raise e [ 571.110572] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.110572] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 571.110572] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.110572] env[63197]: created_port_ids = self._update_ports_for_instance( [ 571.110572] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.110572] env[63197]: with excutils.save_and_reraise_exception(): [ 571.110572] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.110572] env[63197]: self.force_reraise() [ 571.110572] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.110572] env[63197]: raise self.value [ 571.110572] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.110572] env[63197]: updated_port = self._update_port( [ 571.110572] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.110572] env[63197]: _ensure_no_port_binding_failure(port) [ 571.110572] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.110572] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 571.113421] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. [ 571.113421] env[63197]: Removing descriptor: 17 [ 571.113421] env[63197]: ERROR nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Traceback (most recent call last): [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] yield resources [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self.driver.spawn(context, instance, image_meta, [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.113421] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] vm_ref = self.build_virtual_machine(instance, [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] for vif in network_info: [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return self._sync_wrapper(fn, *args, **kwargs) [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self.wait() [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self[:] = self._gt.wait() [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return self._exit_event.wait() [ 571.113927] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] result = hub.switch() [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return self.greenlet.switch() [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] result = function(*args, **kwargs) [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return func(*args, **kwargs) [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] raise e [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] nwinfo = self.network_api.allocate_for_instance( [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.114429] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] created_port_ids = self._update_ports_for_instance( [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] with excutils.save_and_reraise_exception(): [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self.force_reraise() [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] raise self.value [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] updated_port = self._update_port( [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] _ensure_no_port_binding_failure(port) [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.115089] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] raise exception.PortBindingFailed(port_id=port['id']) [ 571.115544] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] nova.exception.PortBindingFailed: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. [ 571.115544] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] [ 571.115544] env[63197]: INFO nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Terminating instance [ 571.115544] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Acquiring lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 571.115544] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Acquired lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 571.115544] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 571.152202] env[63197]: DEBUG nova.compute.utils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.155831] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 571.156027] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 571.333960] env[63197]: DEBUG nova.policy [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f43a1f72c46efa2f36553d1a768ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d97b713fdcb4cd58c6bea06ec23a441', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.415357] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364009, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.02764} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.417969] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] File moved {{(pid=63197) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 571.418439] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Cleaning up location [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 571.418439] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Deleting the datastore file [datastore2] vmware_temp/a185c4fa-8931-4f0a-a7d3-0e28981cc1fd {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 571.418879] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46c89c08-806f-46e1-b0a4-efb5d9f8ac80 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.429279] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 571.429279] env[63197]: value = "task-1364010" [ 571.429279] env[63197]: _type = "Task" [ 571.429279] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.436971] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.658506] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.670738] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.833592] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d17f07d5-3c67-4e5d-869a-40b26479a1ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.843812] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae35b86-1e46-4f4b-9025-2e3e4470da3e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.878963] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b269cf7e-bb3c-4b69-8b29-0e781f14999e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.887559] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c202a36f-119f-4249-9136-0921344b28ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.904305] env[63197]: DEBUG nova.compute.provider_tree [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.939214] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024766} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.940069] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 571.940445] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2fabe8f-d2b5-4e0d-b23d-f1dc2bb7624b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.951282] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 571.951282] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5211d278-e2c1-1c4e-bb75-4a15208bb856" [ 571.951282] env[63197]: _type = "Task" [ 571.951282] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.960964] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5211d278-e2c1-1c4e-bb75-4a15208bb856, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.174706] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.407669] env[63197]: DEBUG nova.scheduler.client.report [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.466858] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5211d278-e2c1-1c4e-bb75-4a15208bb856, 'name': SearchDatastore_Task, 'duration_secs': 0.009166} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.467272] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.467554] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6/9c27469a-b8f9-4374-ae2a-ca17ed43f4f6.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 572.467864] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bfc88963-a588-4468-b023-8a5a58335845 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.476549] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 572.476549] env[63197]: value = "task-1364011" [ 572.476549] env[63197]: _type = "Task" [ 572.476549] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.484780] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364011, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.487106] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Successfully created port: 9cd1d79f-00a0-4402-b99c-6f48616e1579 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.656116] env[63197]: DEBUG nova.compute.manager [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Received event network-vif-deleted-a51c1f11-407a-4fd4-be72-5a92a7d3a456 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.656329] env[63197]: DEBUG nova.compute.manager [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Received event network-changed-8569eb9e-ec1e-4cda-8885-daf63a6ef32e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 572.656619] env[63197]: DEBUG nova.compute.manager [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Refreshing instance network info cache due to event network-changed-8569eb9e-ec1e-4cda-8885-daf63a6ef32e. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 572.656698] env[63197]: DEBUG oslo_concurrency.lockutils [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] Acquiring lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.673072] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.677786] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Releasing lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.678051] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 572.678248] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 572.678809] env[63197]: DEBUG oslo_concurrency.lockutils [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] Acquired lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.678809] env[63197]: DEBUG nova.network.neutron [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Refreshing network info cache for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 572.680482] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6d97fac-41e2-46a3-abe1-4296a103ff8a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.694270] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7e2bd3-3ef2-4fda-901e-71d0ed062153 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.720542] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ec7941d-183d-41d6-b2fb-6e795786c454 could not be found. [ 572.720764] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 572.720944] env[63197]: INFO nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Took 0.04 seconds to destroy the instance on the hypervisor. [ 572.721233] env[63197]: DEBUG oslo.service.loopingcall [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.721493] env[63197]: DEBUG nova.compute.manager [-] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.721590] env[63197]: DEBUG nova.network.neutron [-] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 572.765053] env[63197]: DEBUG nova.network.neutron [-] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.830602] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.831391] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.831721] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.831964] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.832149] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.832470] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.832546] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.832706] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.832880] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.833057] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.833223] env[63197]: DEBUG nova.virt.hardware [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.834557] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80257e85-43aa-495e-ae8c-eefc1efd73b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.846448] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c55b765-80bb-47db-8fab-c6e60d36c0f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.914063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.914636] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 572.919043] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.743s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.919711] env[63197]: INFO nova.compute.claims [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.991995] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364011, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495345} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.992314] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6/9c27469a-b8f9-4374-ae2a-ca17ed43f4f6.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 572.992518] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 572.992760] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a825bbfa-01ef-4ab1-a7e7-66f6225098cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.998589] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 572.998589] env[63197]: value = "task-1364012" [ 572.998589] env[63197]: _type = "Task" [ 572.998589] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.006637] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364012, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.225513] env[63197]: DEBUG nova.network.neutron [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.269155] env[63197]: DEBUG nova.network.neutron [-] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.286237] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Acquiring lock "06de0427-5e1a-469c-ae45-9dc72336d3cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.286466] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Lock "06de0427-5e1a-469c-ae45-9dc72336d3cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.425128] env[63197]: DEBUG nova.compute.utils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.429661] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.429866] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 573.512511] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364012, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.24691} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.513842] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 573.517293] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1ec05e-1f5c-4d84-9ce7-1a32658bfa81 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.521911] env[63197]: DEBUG nova.policy [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f1413ad470d458bbb53556e744946e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '66b7f67f2ef644268d14607908dcdb0b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 573.546963] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6/9c27469a-b8f9-4374-ae2a-ca17ed43f4f6.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 573.547413] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24a493a5-ce81-46f8-b03a-556d9a64e4f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.570349] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 573.570349] env[63197]: value = "task-1364013" [ 573.570349] env[63197]: _type = "Task" [ 573.570349] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.584013] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364013, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.593176] env[63197]: DEBUG nova.network.neutron [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.773229] env[63197]: INFO nova.compute.manager [-] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Took 1.05 seconds to deallocate network for instance. [ 573.775492] env[63197]: DEBUG nova.compute.claims [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 573.775618] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.932412] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.083155] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364013, 'name': ReconfigVM_Task, 'duration_secs': 0.286261} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.085857] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6/9c27469a-b8f9-4374-ae2a-ca17ed43f4f6.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 574.087860] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-061e3ba8-46a7-431d-a5cb-bfc387c9a1fd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.093605] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 574.093605] env[63197]: value = "task-1364014" [ 574.093605] env[63197]: _type = "Task" [ 574.093605] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.097434] env[63197]: DEBUG oslo_concurrency.lockutils [req-d8b2db12-05a6-43e9-a52d-c6e197e60bda req-e8b31109-0cd9-43f6-bcaa-726e41077e8b service nova] Releasing lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.103501] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364014, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.157020] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a4ff7d-80d9-473f-8e75-61fc01ecb6f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.164607] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcb41b6-0a34-4c7e-a9ad-38177acf8744 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.205553] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb4d6b4-e1e6-4294-a2c5-92541146a887 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.212846] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef114fb-b851-44d6-b1bd-15881008b1c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.227018] env[63197]: DEBUG nova.compute.provider_tree [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.373116] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Acquiring lock "b6581d2d-51e8-431f-97d6-acda6f438e9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.373116] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Lock "b6581d2d-51e8-431f-97d6-acda6f438e9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.563248] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Successfully created port: 0042be2f-088c-4b82-a376-3dfabacf34fd {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.605562] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364014, 'name': Rename_Task, 'duration_secs': 0.177597} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.605562] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 574.605966] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6b0b452-64b6-4725-b888-63cfda76a3aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.615922] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 574.615922] env[63197]: value = "task-1364015" [ 574.615922] env[63197]: _type = "Task" [ 574.615922] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.626793] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364015, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.733505] env[63197]: DEBUG nova.scheduler.client.report [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.942820] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Acquiring lock "2d8fb706-7400-43af-86fe-f90ac38393e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.942820] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Lock "2d8fb706-7400-43af-86fe-f90ac38393e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.949035] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 574.977360] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.977360] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.977360] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.977650] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.977650] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.977650] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.977650] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.977650] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.977816] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.977816] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.977816] env[63197]: DEBUG nova.virt.hardware [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.978143] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9556e05-7e3a-4409-a202-66a66cf6763a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.987183] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d7e9d0-9205-46ee-b19a-24cfec4bef9a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.134289] env[63197]: DEBUG oslo_vmware.api [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364015, 'name': PowerOnVM_Task, 'duration_secs': 0.456259} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.134658] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 575.135033] env[63197]: INFO nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Took 9.61 seconds to spawn the instance on the hypervisor. [ 575.135536] env[63197]: DEBUG nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 575.136565] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd8aac5-7bca-4273-937e-333f6ad213f8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.244765] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.245303] env[63197]: DEBUG nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.248206] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.721s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.250886] env[63197]: INFO nova.compute.claims [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.319517] env[63197]: ERROR nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. [ 575.319517] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.319517] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.319517] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.319517] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.319517] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.319517] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.319517] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.319517] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.319517] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 575.319517] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.319517] env[63197]: ERROR nova.compute.manager raise self.value [ 575.319517] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.319517] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.319517] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.319517] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.319955] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.319955] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.319955] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. [ 575.319955] env[63197]: ERROR nova.compute.manager [ 575.319955] env[63197]: Traceback (most recent call last): [ 575.319955] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.319955] env[63197]: listener.cb(fileno) [ 575.319955] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.319955] env[63197]: result = function(*args, **kwargs) [ 575.319955] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.319955] env[63197]: return func(*args, **kwargs) [ 575.319955] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.319955] env[63197]: raise e [ 575.319955] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.319955] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 575.319955] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.319955] env[63197]: created_port_ids = self._update_ports_for_instance( [ 575.319955] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.319955] env[63197]: with excutils.save_and_reraise_exception(): [ 575.319955] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.319955] env[63197]: self.force_reraise() [ 575.319955] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.319955] env[63197]: raise self.value [ 575.319955] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.319955] env[63197]: updated_port = self._update_port( [ 575.319955] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.319955] env[63197]: _ensure_no_port_binding_failure(port) [ 575.319955] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.319955] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.320743] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. [ 575.320743] env[63197]: Removing descriptor: 15 [ 575.320743] env[63197]: ERROR nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Traceback (most recent call last): [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] yield resources [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self.driver.spawn(context, instance, image_meta, [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.320743] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] vm_ref = self.build_virtual_machine(instance, [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] for vif in network_info: [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return self._sync_wrapper(fn, *args, **kwargs) [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self.wait() [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self[:] = self._gt.wait() [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return self._exit_event.wait() [ 575.321170] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] result = hub.switch() [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return self.greenlet.switch() [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] result = function(*args, **kwargs) [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return func(*args, **kwargs) [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] raise e [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] nwinfo = self.network_api.allocate_for_instance( [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.321522] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] created_port_ids = self._update_ports_for_instance( [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] with excutils.save_and_reraise_exception(): [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self.force_reraise() [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] raise self.value [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] updated_port = self._update_port( [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] _ensure_no_port_binding_failure(port) [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.321840] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] raise exception.PortBindingFailed(port_id=port['id']) [ 575.322153] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] nova.exception.PortBindingFailed: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. [ 575.322153] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] [ 575.322153] env[63197]: INFO nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Terminating instance [ 575.322153] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Acquiring lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.322153] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Acquired lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.322153] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 575.661625] env[63197]: INFO nova.compute.manager [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Took 15.15 seconds to build instance. [ 575.758133] env[63197]: DEBUG nova.compute.utils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.762320] env[63197]: DEBUG nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Not allocating networking since 'none' was specified. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 575.895092] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.166207] env[63197]: DEBUG oslo_concurrency.lockutils [None req-560afa7f-08a7-4379-bda6-0a89d2ef19a5 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.675s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.263428] env[63197]: DEBUG nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.393516] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.486022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Acquiring lock "441988d7-b5c1-4996-a6fa-4a6e21f86b22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.486022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Lock "441988d7-b5c1-4996-a6fa-4a6e21f86b22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.487280] env[63197]: DEBUG nova.compute.manager [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Received event network-vif-deleted-8569eb9e-ec1e-4cda-8885-daf63a6ef32e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.487553] env[63197]: DEBUG nova.compute.manager [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Received event network-changed-9d17dab0-f776-4eb7-8729-0ede74a50ddd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.488046] env[63197]: DEBUG nova.compute.manager [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Refreshing instance network info cache due to event network-changed-9d17dab0-f776-4eb7-8729-0ede74a50ddd. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 576.488046] env[63197]: DEBUG oslo_concurrency.lockutils [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] Acquiring lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.525974] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f93e213-9757-4164-af47-ef0647c1bc2c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.535528] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f8c2ca-7b82-4049-962a-a331172502a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.577383] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f096fce0-5d34-4134-ad0e-f80d844adb62 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.584776] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b47214-626b-4ae9-9aaa-197b0ccb16fa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.601886] env[63197]: DEBUG nova.compute.provider_tree [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.668717] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 576.895976] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Releasing lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.899075] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 576.899075] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 576.899075] env[63197]: DEBUG oslo_concurrency.lockutils [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] Acquired lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.899075] env[63197]: DEBUG nova.network.neutron [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Refreshing network info cache for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 576.899075] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-679f1ecd-2201-40ab-8f89-32aabae702db {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.910521] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200e42be-7f6f-4909-bda0-7c9e9f804e26 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.939299] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a892c650-d8e8-4eaa-b025-cf330478aabf could not be found. [ 576.939968] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 576.940368] env[63197]: INFO nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 576.940755] env[63197]: DEBUG oslo.service.loopingcall [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.941128] env[63197]: DEBUG nova.compute.manager [-] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.941397] env[63197]: DEBUG nova.network.neutron [-] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 576.996898] env[63197]: DEBUG nova.network.neutron [-] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.108067] env[63197]: DEBUG nova.scheduler.client.report [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.195089] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.277545] env[63197]: DEBUG nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.302256] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.302549] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.302738] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.302952] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.303314] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.303528] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.303779] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.303998] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.304355] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.304879] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.305124] env[63197]: DEBUG nova.virt.hardware [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.306030] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e932d8ec-af47-40c4-918c-0f469fad31ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.314458] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa696848-3f02-4e7d-b12f-fa48f9495ec7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.330909] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 577.338242] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Creating folder: Project (39188f716c3245e6b66b71e57ebf0af1). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 577.338606] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4fea0322-b60f-4455-85d2-b26b0d9d49d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.348536] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Created folder: Project (39188f716c3245e6b66b71e57ebf0af1) in parent group-v290286. [ 577.348730] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Creating folder: Instances. Parent ref: group-v290290. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 577.348981] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad27fe4e-1ef9-4d77-af1e-c3f16606dd87 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.358026] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Created folder: Instances in parent group-v290290. [ 577.358295] env[63197]: DEBUG oslo.service.loopingcall [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.358562] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 577.358682] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65f195e7-8d8e-435c-8e2d-1621ac7b3392 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.375924] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 577.375924] env[63197]: value = "task-1364018" [ 577.375924] env[63197]: _type = "Task" [ 577.375924] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.383628] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364018, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.482495] env[63197]: DEBUG nova.network.neutron [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.498893] env[63197]: DEBUG nova.network.neutron [-] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.613014] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.613574] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.616941] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.767s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.620232] env[63197]: INFO nova.compute.claims [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.890360] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364018, 'name': CreateVM_Task, 'duration_secs': 0.313235} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.890547] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 577.891461] env[63197]: DEBUG oslo_vmware.service [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b541b05-02ad-4f2c-8e9b-164802bd99e1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.897735] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.897894] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.898289] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 577.898565] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f291e896-6168-42fb-b9d5-8d7b40df2488 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.904389] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 577.904389] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b504ba-28f5-91c5-5e73-6d4698d5fcfa" [ 577.904389] env[63197]: _type = "Task" [ 577.904389] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.913340] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b504ba-28f5-91c5-5e73-6d4698d5fcfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.003449] env[63197]: INFO nova.compute.manager [-] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Took 1.06 seconds to deallocate network for instance. [ 578.006314] env[63197]: DEBUG nova.compute.claims [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.008198] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.123099] env[63197]: DEBUG nova.compute.utils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.126652] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.126947] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 578.140811] env[63197]: DEBUG nova.compute.manager [None req-9b3bdf0a-1b01-4b7f-8a42-50d1a11b6402 tempest-ServerDiagnosticsV248Test-400240072 tempest-ServerDiagnosticsV248Test-400240072-project-admin] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 578.143311] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5370b7c-28d6-48da-8a1a-3d65ff6decd6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.155013] env[63197]: INFO nova.compute.manager [None req-9b3bdf0a-1b01-4b7f-8a42-50d1a11b6402 tempest-ServerDiagnosticsV248Test-400240072 tempest-ServerDiagnosticsV248Test-400240072-project-admin] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Retrieving diagnostics [ 578.155931] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448e80f5-5d57-4428-9c98-a5be289454de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.195810] env[63197]: DEBUG nova.network.neutron [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.209397] env[63197]: ERROR nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. [ 578.209397] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.209397] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.209397] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.209397] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.209397] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.209397] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.209397] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.209397] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.209397] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 578.209397] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.209397] env[63197]: ERROR nova.compute.manager raise self.value [ 578.209397] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.209397] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.209397] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.209397] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.209888] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.209888] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.209888] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. [ 578.209888] env[63197]: ERROR nova.compute.manager [ 578.209888] env[63197]: Traceback (most recent call last): [ 578.209888] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.209888] env[63197]: listener.cb(fileno) [ 578.209888] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.209888] env[63197]: result = function(*args, **kwargs) [ 578.209888] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.209888] env[63197]: return func(*args, **kwargs) [ 578.209888] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.209888] env[63197]: raise e [ 578.209888] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.209888] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 578.209888] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.209888] env[63197]: created_port_ids = self._update_ports_for_instance( [ 578.209888] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.209888] env[63197]: with excutils.save_and_reraise_exception(): [ 578.209888] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.209888] env[63197]: self.force_reraise() [ 578.209888] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.209888] env[63197]: raise self.value [ 578.209888] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.209888] env[63197]: updated_port = self._update_port( [ 578.209888] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.209888] env[63197]: _ensure_no_port_binding_failure(port) [ 578.209888] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.209888] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.210775] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. [ 578.210775] env[63197]: Removing descriptor: 19 [ 578.210775] env[63197]: ERROR nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Traceback (most recent call last): [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] yield resources [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self.driver.spawn(context, instance, image_meta, [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.210775] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] vm_ref = self.build_virtual_machine(instance, [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] for vif in network_info: [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return self._sync_wrapper(fn, *args, **kwargs) [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self.wait() [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self[:] = self._gt.wait() [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return self._exit_event.wait() [ 578.211179] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] result = hub.switch() [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return self.greenlet.switch() [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] result = function(*args, **kwargs) [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return func(*args, **kwargs) [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] raise e [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] nwinfo = self.network_api.allocate_for_instance( [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.211573] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] created_port_ids = self._update_ports_for_instance( [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] with excutils.save_and_reraise_exception(): [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self.force_reraise() [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] raise self.value [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] updated_port = self._update_port( [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] _ensure_no_port_binding_failure(port) [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.211970] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] raise exception.PortBindingFailed(port_id=port['id']) [ 578.212341] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] nova.exception.PortBindingFailed: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. [ 578.212341] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] [ 578.212341] env[63197]: INFO nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Terminating instance [ 578.213764] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Acquiring lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.213764] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Acquired lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.213764] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.282019] env[63197]: ERROR nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. [ 578.282019] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.282019] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.282019] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.282019] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.282019] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.282019] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.282019] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.282019] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.282019] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 578.282019] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.282019] env[63197]: ERROR nova.compute.manager raise self.value [ 578.282019] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.282019] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.282019] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.282019] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.282576] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.282576] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.282576] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. [ 578.282576] env[63197]: ERROR nova.compute.manager [ 578.282576] env[63197]: Traceback (most recent call last): [ 578.282576] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.282576] env[63197]: listener.cb(fileno) [ 578.282576] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.282576] env[63197]: result = function(*args, **kwargs) [ 578.282576] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.282576] env[63197]: return func(*args, **kwargs) [ 578.282576] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.282576] env[63197]: raise e [ 578.282576] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.282576] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 578.282576] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.282576] env[63197]: created_port_ids = self._update_ports_for_instance( [ 578.282576] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.282576] env[63197]: with excutils.save_and_reraise_exception(): [ 578.282576] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.282576] env[63197]: self.force_reraise() [ 578.282576] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.282576] env[63197]: raise self.value [ 578.282576] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.282576] env[63197]: updated_port = self._update_port( [ 578.282576] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.282576] env[63197]: _ensure_no_port_binding_failure(port) [ 578.282576] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.282576] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.283405] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. [ 578.283405] env[63197]: Removing descriptor: 17 [ 578.283405] env[63197]: ERROR nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Traceback (most recent call last): [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] yield resources [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self.driver.spawn(context, instance, image_meta, [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.283405] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] vm_ref = self.build_virtual_machine(instance, [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] for vif in network_info: [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return self._sync_wrapper(fn, *args, **kwargs) [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self.wait() [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self[:] = self._gt.wait() [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return self._exit_event.wait() [ 578.283759] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] result = hub.switch() [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return self.greenlet.switch() [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] result = function(*args, **kwargs) [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return func(*args, **kwargs) [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] raise e [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] nwinfo = self.network_api.allocate_for_instance( [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.284081] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] created_port_ids = self._update_ports_for_instance( [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] with excutils.save_and_reraise_exception(): [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self.force_reraise() [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] raise self.value [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] updated_port = self._update_port( [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] _ensure_no_port_binding_failure(port) [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.284409] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] raise exception.PortBindingFailed(port_id=port['id']) [ 578.284754] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] nova.exception.PortBindingFailed: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. [ 578.284754] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] [ 578.284754] env[63197]: INFO nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Terminating instance [ 578.284754] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.284754] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.284754] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.383445] env[63197]: DEBUG nova.policy [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b43a04053e341f1a7032aa58befb46a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5777c4b4ccad4dadb0e5a93c5a5b3231', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.417652] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.417904] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 578.418152] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.418300] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.418473] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 578.418720] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96e3b92a-14aa-44fc-a755-9ce74cc0fa70 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.427220] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 578.427697] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 578.428507] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80151915-f793-4da2-825f-dcc95660dabe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.435891] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57784c55-fc8f-4022-b27f-1ed101696dea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.440617] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 578.440617] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528510d0-7e45-420b-9e4a-58d367b887c1" [ 578.440617] env[63197]: _type = "Task" [ 578.440617] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.447935] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528510d0-7e45-420b-9e4a-58d367b887c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.622093] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "bb7f3278-8678-4053-b957-9ceee430089d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.622809] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "bb7f3278-8678-4053-b957-9ceee430089d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.627477] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.699796] env[63197]: DEBUG oslo_concurrency.lockutils [req-1aa5ebc5-c553-4670-af77-e4b488e63169 req-57a7728e-a0d4-4642-970f-8a87f7ac9be6 service nova] Releasing lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.776544] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.924686] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef63cb2-93d8-4040-8a62-1ed8812b925a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.931074] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.937255] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2882698a-6c73-4ab7-9295-57bb7647930e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.976194] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1162ff7e-336f-45d5-b5e5-e63a5c50f980 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.983217] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Preparing fetch location {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 578.983728] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Creating directory with path [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 578.984084] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d2b808e-8d9f-4a18-91e0-6cc1013734b7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.990637] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573b24e9-9d01-45e9-b3b3-3615bcaf4478 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.005172] env[63197]: DEBUG nova.compute.provider_tree [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.007880] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.012421] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Created directory with path [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 579.012421] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Fetch image to [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 579.012421] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Downloading image file data f760fedc-0b5b-4c56-acbe-239b47c945e8 to [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk on the data store datastore1 {{(pid=63197) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 579.012421] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243648e4-723d-447a-9526-8685da048f66 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.018778] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44b7d23-fa32-4d8b-bd37-8d1214ad0a3a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.033472] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b18644f-7a14-411b-a9ba-e8f068dba626 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.066038] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8896118-fdec-43d5-b62f-9b9685bcf6a3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.072119] env[63197]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-2a92738d-d486-4485-8d0e-07d4d4571278 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.159983] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Downloading image file data f760fedc-0b5b-4c56-acbe-239b47c945e8 to the data store datastore1 {{(pid=63197) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 579.197871] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.240035] env[63197]: DEBUG oslo_vmware.rw_handles [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63197) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 579.306058] env[63197]: DEBUG nova.compute.manager [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Received event network-changed-0042be2f-088c-4b82-a376-3dfabacf34fd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.306058] env[63197]: DEBUG nova.compute.manager [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Refreshing instance network info cache due to event network-changed-0042be2f-088c-4b82-a376-3dfabacf34fd. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.306058] env[63197]: DEBUG oslo_concurrency.lockutils [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] Acquiring lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.515022] env[63197]: DEBUG nova.scheduler.client.report [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.518471] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Successfully created port: 56564457-bbf2-48f2-af44-708898339d9f {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.521670] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Releasing lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.522126] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.522369] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.523104] env[63197]: DEBUG oslo_concurrency.lockutils [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] Acquired lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.523288] env[63197]: DEBUG nova.network.neutron [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Refreshing network info cache for port 0042be2f-088c-4b82-a376-3dfabacf34fd {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 579.525350] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6079ef6-a82f-400f-a8ef-7192c1a37276 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.538427] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 579.538427] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 579.538427] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 579.538427] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Rebuilding the list of instances to heal {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 579.545656] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8261ae-0a23-4e91-ba3e-53e14fca8e80 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.589107] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c0742ebb-c176-44cf-ae13-d778cc6364cb could not be found. [ 579.589107] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.589107] env[63197]: INFO nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Took 0.07 seconds to destroy the instance on the hypervisor. [ 579.589585] env[63197]: DEBUG oslo.service.loopingcall [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.593648] env[63197]: DEBUG nova.compute.manager [-] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.593978] env[63197]: DEBUG nova.network.neutron [-] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.637400] env[63197]: DEBUG nova.network.neutron [-] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.646296] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.691376] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.691668] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.691969] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.692074] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.692262] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.692428] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.692687] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.692827] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.693553] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.693553] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.693553] env[63197]: DEBUG nova.virt.hardware [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.694744] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0643a3-eb65-4856-a3e6-6d22e4e28a9d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.700659] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.701666] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 579.701666] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 579.701906] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-365f280d-831c-448a-9cb9-93050568431b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.714831] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f8d902-6c5e-46a3-a4a0-fd6264ee454d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.732771] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a380d821-c231-4d60-bd48-9d4d3383168f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.771786] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e could not be found. [ 579.771786] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 579.771786] env[63197]: INFO nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Took 0.07 seconds to destroy the instance on the hypervisor. [ 579.771786] env[63197]: DEBUG oslo.service.loopingcall [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 579.771786] env[63197]: DEBUG nova.compute.manager [-] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.771786] env[63197]: DEBUG nova.network.neutron [-] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.812363] env[63197]: DEBUG nova.network.neutron [-] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.901087] env[63197]: DEBUG oslo_vmware.rw_handles [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Completed reading data from the image iterator. {{(pid=63197) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 579.901331] env[63197]: DEBUG oslo_vmware.rw_handles [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 580.024858] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.025078] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.027654] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.086s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.041852] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.042028] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.042158] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.042280] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.042395] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.042510] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.042623] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.042737] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 580.047864] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Downloaded image file data f760fedc-0b5b-4c56-acbe-239b47c945e8 to vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk on the data store datastore1 {{(pid=63197) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 580.049679] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Caching image {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 580.049914] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Copying Virtual Disk [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk to [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 580.050224] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6264b1d0-41ed-49ed-8893-5ffdd1c5fe55 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.063867] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 580.063867] env[63197]: value = "task-1364019" [ 580.063867] env[63197]: _type = "Task" [ 580.063867] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.072437] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364019, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.114119] env[63197]: DEBUG nova.network.neutron [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.140871] env[63197]: DEBUG nova.network.neutron [-] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.153182] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "refresh_cache-9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.153386] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquired lock "refresh_cache-9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.153534] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Forcefully refreshing network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 580.153753] env[63197]: DEBUG nova.objects.instance [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lazy-loading 'info_cache' on Instance uuid 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 580.316249] env[63197]: DEBUG nova.compute.manager [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Received event network-vif-deleted-9d17dab0-f776-4eb7-8729-0ede74a50ddd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.316522] env[63197]: DEBUG nova.compute.manager [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Received event network-changed-9cd1d79f-00a0-4402-b99c-6f48616e1579 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.316823] env[63197]: DEBUG nova.compute.manager [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Refreshing instance network info cache due to event network-changed-9cd1d79f-00a0-4402-b99c-6f48616e1579. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 580.317327] env[63197]: DEBUG oslo_concurrency.lockutils [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] Acquiring lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.317498] env[63197]: DEBUG oslo_concurrency.lockutils [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] Acquired lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.317717] env[63197]: DEBUG nova.network.neutron [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Refreshing network info cache for port 9cd1d79f-00a0-4402-b99c-6f48616e1579 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.322932] env[63197]: DEBUG nova.network.neutron [-] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.331748] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Acquiring lock "5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.331748] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Lock "5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.435697] env[63197]: DEBUG nova.network.neutron [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.533548] env[63197]: DEBUG nova.compute.utils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.535198] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.535429] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 580.577456] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364019, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.643454] env[63197]: INFO nova.compute.manager [-] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Took 1.05 seconds to deallocate network for instance. [ 580.650387] env[63197]: DEBUG nova.compute.claims [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.650387] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.816444] env[63197]: DEBUG nova.policy [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3356912d43b451ea8cfd31abe9fbced', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae77ae8a2fea4217b5620b7e325298e8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.825174] env[63197]: INFO nova.compute.manager [-] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Took 1.05 seconds to deallocate network for instance. [ 580.829553] env[63197]: DEBUG nova.compute.claims [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 580.829732] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.903233] env[63197]: DEBUG nova.network.neutron [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.939661] env[63197]: DEBUG oslo_concurrency.lockutils [req-7a1b0dce-a24a-476a-ae04-75ebee181e3e req-c584a47c-434e-4e83-90f8-ccecbb94eb23 service nova] Releasing lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.955960] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16a9a55-619f-4412-82f5-86b456fe4c15 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.967559] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51c2f65-50bf-47e0-8328-d1c115412f6b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.009720] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4ea2ec-9545-4649-b369-16986bbdfc24 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.016469] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fb493d-ca3a-45ef-af23-15556daa5e51 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.032044] env[63197]: DEBUG nova.compute.provider_tree [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.043898] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.082609] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364019, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.708997} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.082609] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Copied Virtual Disk [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk to [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 581.082609] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleting the datastore file [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8/tmp-sparse.vmdk {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 581.085224] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfdaa4cf-1c42-4595-b248-e6dee20d110b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.095224] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 581.095224] env[63197]: value = "task-1364020" [ 581.095224] env[63197]: _type = "Task" [ 581.095224] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.106459] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364020, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.221706] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.318353] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "d84f11a3-ed86-4fc4-9092-0babc7fccb3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.318929] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "d84f11a3-ed86-4fc4-9092-0babc7fccb3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 581.319523] env[63197]: DEBUG nova.network.neutron [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.535782] env[63197]: DEBUG nova.scheduler.client.report [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.610825] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364020, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022185} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.611254] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 581.611488] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Moving file from [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7/f760fedc-0b5b-4c56-acbe-239b47c945e8 to [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8. {{(pid=63197) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 581.611748] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-ec4952d9-466d-4089-b6de-ae9c40e7cd74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.619521] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 581.619521] env[63197]: value = "task-1364021" [ 581.619521] env[63197]: _type = "Task" [ 581.619521] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.627881] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364021, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.822835] env[63197]: DEBUG oslo_concurrency.lockutils [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] Releasing lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.823144] env[63197]: DEBUG nova.compute.manager [req-39210323-0e77-4e7c-a762-1317d3dadea2 req-3c32f838-2f24-43d3-be85-56c42f18731d service nova] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Received event network-vif-deleted-9cd1d79f-00a0-4402-b99c-6f48616e1579 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.015874] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.042350] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.042825] env[63197]: ERROR nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Traceback (most recent call last): [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self.driver.spawn(context, instance, image_meta, [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] vm_ref = self.build_virtual_machine(instance, [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.042825] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] for vif in network_info: [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return self._sync_wrapper(fn, *args, **kwargs) [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self.wait() [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self[:] = self._gt.wait() [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return self._exit_event.wait() [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] result = hub.switch() [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.043149] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return self.greenlet.switch() [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] result = function(*args, **kwargs) [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] return func(*args, **kwargs) [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] raise e [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] nwinfo = self.network_api.allocate_for_instance( [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] created_port_ids = self._update_ports_for_instance( [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] with excutils.save_and_reraise_exception(): [ 582.043477] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] self.force_reraise() [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] raise self.value [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] updated_port = self._update_port( [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] _ensure_no_port_binding_failure(port) [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] raise exception.PortBindingFailed(port_id=port['id']) [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] nova.exception.PortBindingFailed: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. [ 582.043796] env[63197]: ERROR nova.compute.manager [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] [ 582.044116] env[63197]: DEBUG nova.compute.utils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.045163] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.819s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.046702] env[63197]: INFO nova.compute.claims [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.059437] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.059437] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Build of instance ac9d308b-ac4b-4153-8e1d-7551213c8233 was re-scheduled: Binding failed for port a51c1f11-407a-4fd4-be72-5a92a7d3a456, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.060032] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.062224] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.062224] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquired lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.062224] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.098724] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.099019] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.099242] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.100661] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.100661] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.100661] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.100661] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.100661] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.101267] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.101267] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.101338] env[63197]: DEBUG nova.virt.hardware [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.105047] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b36c6c4-c33a-4bda-8805-626cdd57aab6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.112678] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7c7006-7cf5-400a-a675-eb9ff59909c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.138828] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364021, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023028} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.139144] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] File moved {{(pid=63197) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 582.139358] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Cleaning up location [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 582.139522] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleting the datastore file [datastore1] vmware_temp/f2fedef4-3c02-4fab-bdd5-6637e65adde7 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 582.139850] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94735096-6272-4841-8a84-90420dad4b3f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.150479] env[63197]: DEBUG nova.compute.manager [req-186bb9c7-1633-4fc8-9740-63a9cf842a57 req-f8d4f094-1a94-4918-a94f-958f27daec5f service nova] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Received event network-vif-deleted-0042be2f-088c-4b82-a376-3dfabacf34fd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.153759] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 582.153759] env[63197]: value = "task-1364022" [ 582.153759] env[63197]: _type = "Task" [ 582.153759] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.165106] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364022, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.519848] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Releasing lock "refresh_cache-9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.520153] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Updated the network info_cache for instance {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 582.520153] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 582.520442] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 582.520644] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 582.521625] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 582.521625] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 582.521625] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._sync_power_states {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 582.609790] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.664897] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364022, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025005} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.665361] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 582.665918] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b2ce054-25cd-47b5-9a0b-a0095cab3b20 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.672068] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 582.672068] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52040b72-787a-d82c-3948-7bd64beb7df3" [ 582.672068] env[63197]: _type = "Task" [ 582.672068] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.682766] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Successfully created port: e312e7b2-5270-4cf6-8678-7a1af450ad88 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.688586] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52040b72-787a-d82c-3948-7bd64beb7df3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.838377] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.025594] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Getting list of instances from cluster (obj){ [ 583.025594] env[63197]: value = "domain-c8" [ 583.025594] env[63197]: _type = "ClusterComputeResource" [ 583.025594] env[63197]: } {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 583.027433] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2f6205-aba4-4edb-9a43-412dd788b66a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.048548] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Got total of 2 instances {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 583.048548] env[63197]: WARNING nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] While synchronizing instance power states, found 9 instances in the database and 2 instances on the hypervisor. [ 583.048548] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6 {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.048548] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid 7ec7941d-183d-41d6-b2fb-6e795786c454 {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.048548] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid a892c650-d8e8-4eaa-b025-cf330478aabf {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.048980] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.049295] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid c0742ebb-c176-44cf-ae13-d778cc6364cb {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.049794] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid 99539d7f-cd93-4250-97cd-2449e03e517d {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.050801] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid 68ce48d7-e967-4cfc-a607-f6580bc14648 {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.050801] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid 37280bd4-3cd8-4910-97de-0cc7494abc22 {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.051468] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid a8315374-abb6-4663-bdb5-c4d88cdbbaa4 {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 583.052149] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.052510] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.052995] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "7ec7941d-183d-41d6-b2fb-6e795786c454" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.053352] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "a892c650-d8e8-4eaa-b025-cf330478aabf" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.053670] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.056019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "c0742ebb-c176-44cf-ae13-d778cc6364cb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.056019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "99539d7f-cd93-4250-97cd-2449e03e517d" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.056019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "68ce48d7-e967-4cfc-a607-f6580bc14648" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.056019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "37280bd4-3cd8-4910-97de-0cc7494abc22" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.056019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "a8315374-abb6-4663-bdb5-c4d88cdbbaa4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.056341] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 583.056341] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 583.057048] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b23ef6c-f787-49d5-b215-7115613632ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.061193] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 583.185129] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52040b72-787a-d82c-3948-7bd64beb7df3, 'name': SearchDatastore_Task, 'duration_secs': 0.010969} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.189655] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.189655] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 583.189655] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90272dea-aa0f-4b9c-9ea7-65e8193d8fcf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.198650] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 583.198650] env[63197]: value = "task-1364023" [ 583.198650] env[63197]: _type = "Task" [ 583.198650] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.210439] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.342167] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Releasing lock "refresh_cache-ac9d308b-ac4b-4153-8e1d-7551213c8233" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.344020] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.344020] env[63197]: DEBUG nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.344020] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.388774] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3ee8bb-9e48-4f2b-85d1-f81a731f369d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.398216] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959bd9cc-7079-40b1-99f2-fcaa6e06c2ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.402466] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.444567] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6130a175-76aa-41c1-ae37-65ade77a6a7e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.454183] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd01284-afdf-4e16-a3c4-7b5082c93ed3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.473298] env[63197]: DEBUG nova.compute.provider_tree [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.570014] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.577097] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.524s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.716301] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364023, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509522} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.716709] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 583.716709] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 583.718348] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50de81da-5de6-4e7f-94ac-8172d3c89323 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.724335] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 583.724335] env[63197]: value = "task-1364024" [ 583.724335] env[63197]: _type = "Task" [ 583.724335] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.733036] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364024, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.886921] env[63197]: DEBUG nova.compute.manager [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Received event network-changed-56564457-bbf2-48f2-af44-708898339d9f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.887128] env[63197]: DEBUG nova.compute.manager [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Refreshing instance network info cache due to event network-changed-56564457-bbf2-48f2-af44-708898339d9f. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 583.887344] env[63197]: DEBUG oslo_concurrency.lockutils [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] Acquiring lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.887513] env[63197]: DEBUG oslo_concurrency.lockutils [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] Acquired lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.887632] env[63197]: DEBUG nova.network.neutron [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Refreshing network info cache for port 56564457-bbf2-48f2-af44-708898339d9f {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.893643] env[63197]: ERROR nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. [ 583.893643] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 583.893643] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.893643] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 583.893643] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.893643] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 583.893643] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.893643] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 583.893643] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.893643] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 583.893643] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.893643] env[63197]: ERROR nova.compute.manager raise self.value [ 583.893643] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.893643] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 583.893643] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.893643] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 583.894286] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.894286] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 583.894286] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. [ 583.894286] env[63197]: ERROR nova.compute.manager [ 583.894286] env[63197]: Traceback (most recent call last): [ 583.894286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 583.894286] env[63197]: listener.cb(fileno) [ 583.894286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.894286] env[63197]: result = function(*args, **kwargs) [ 583.894286] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.894286] env[63197]: return func(*args, **kwargs) [ 583.894286] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.894286] env[63197]: raise e [ 583.894286] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.894286] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 583.894286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.894286] env[63197]: created_port_ids = self._update_ports_for_instance( [ 583.894286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.894286] env[63197]: with excutils.save_and_reraise_exception(): [ 583.894286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.894286] env[63197]: self.force_reraise() [ 583.894286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.894286] env[63197]: raise self.value [ 583.894286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.894286] env[63197]: updated_port = self._update_port( [ 583.894286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.894286] env[63197]: _ensure_no_port_binding_failure(port) [ 583.894286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.894286] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 583.895679] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. [ 583.895679] env[63197]: Removing descriptor: 18 [ 583.895679] env[63197]: ERROR nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Traceback (most recent call last): [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] yield resources [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self.driver.spawn(context, instance, image_meta, [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.895679] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] vm_ref = self.build_virtual_machine(instance, [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] for vif in network_info: [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return self._sync_wrapper(fn, *args, **kwargs) [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self.wait() [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self[:] = self._gt.wait() [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return self._exit_event.wait() [ 583.898285] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] result = hub.switch() [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return self.greenlet.switch() [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] result = function(*args, **kwargs) [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return func(*args, **kwargs) [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] raise e [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] nwinfo = self.network_api.allocate_for_instance( [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.898955] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] created_port_ids = self._update_ports_for_instance( [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] with excutils.save_and_reraise_exception(): [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self.force_reraise() [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] raise self.value [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] updated_port = self._update_port( [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] _ensure_no_port_binding_failure(port) [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.899348] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] raise exception.PortBindingFailed(port_id=port['id']) [ 583.899698] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] nova.exception.PortBindingFailed: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. [ 583.899698] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] [ 583.899698] env[63197]: INFO nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Terminating instance [ 583.901213] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Acquiring lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.910090] env[63197]: DEBUG nova.network.neutron [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.976741] env[63197]: DEBUG nova.scheduler.client.report [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.239376] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364024, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063562} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.239574] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 584.240593] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b803d649-481c-4bf9-9670-8ec289768c8e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.271213] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 584.271849] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2dc019e-9854-4cee-ba96-57c88a09dcc9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.294795] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 584.294795] env[63197]: value = "task-1364025" [ 584.294795] env[63197]: _type = "Task" [ 584.294795] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.303111] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364025, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.411728] env[63197]: INFO nova.compute.manager [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: ac9d308b-ac4b-4153-8e1d-7551213c8233] Took 1.07 seconds to deallocate network for instance. [ 584.440677] env[63197]: DEBUG nova.network.neutron [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.489501] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.490016] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 584.493838] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.718s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.627535] env[63197]: DEBUG nova.network.neutron [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.809060] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364025, 'name': ReconfigVM_Task, 'duration_secs': 0.314092} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.811695] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Reconfigured VM instance instance-00000007 to attach disk [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 584.812818] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ffcb23da-e2cb-473f-ac2a-32e5b646ffd8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.822080] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 584.822080] env[63197]: value = "task-1364026" [ 584.822080] env[63197]: _type = "Task" [ 584.822080] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.832607] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364026, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.924732] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.928355] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.996198] env[63197]: DEBUG nova.compute.utils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 584.998044] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 584.998044] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.133712] env[63197]: DEBUG oslo_concurrency.lockutils [req-6442001a-6d69-4c9c-a11f-cde0ce5b052f req-ba4aedd7-0edb-46a0-93a6-ab8d9f9d79cb service nova] Releasing lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.135544] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Acquired lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.135544] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.192484] env[63197]: DEBUG nova.policy [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45d0ddb8ab2e463e9b90ff6fd3c16a20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcfb3f4823944481a80cc9fb92209ca7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.331978] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364026, 'name': Rename_Task, 'duration_secs': 0.13293} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.336761] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 585.337984] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80f2f8be-b8ad-4c4b-8952-ff334a8b205f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.345206] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 585.345206] env[63197]: value = "task-1364027" [ 585.345206] env[63197]: _type = "Task" [ 585.345206] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.352322] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129ec8f0-d217-44ca-a073-e6abab0e7e4c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.360196] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364027, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.362618] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c565e094-fe0f-4145-b283-eaff863a5b89 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.396776] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2fbda8-8680-4e5d-ba30-a4f030ba8169 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.404374] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce379a4d-cb0b-4ba0-b5cf-ad40321c465b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.420677] env[63197]: DEBUG nova.compute.provider_tree [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.459347] env[63197]: INFO nova.scheduler.client.report [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Deleted allocations for instance ac9d308b-ac4b-4153-8e1d-7551213c8233 [ 585.500837] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 585.684378] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.856323] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Acquiring lock "f01a025d-2cb9-4cb9-878a-b05e318c21b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.856568] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Lock "f01a025d-2cb9-4cb9-878a-b05e318c21b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.863337] env[63197]: DEBUG oslo_vmware.api [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364027, 'name': PowerOnVM_Task, 'duration_secs': 0.497881} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.863575] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 585.863752] env[63197]: INFO nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Took 8.59 seconds to spawn the instance on the hypervisor. [ 585.863945] env[63197]: DEBUG nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 585.864731] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3548751a-2d60-4920-85ab-1eb6cc41c0bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.927263] env[63197]: DEBUG nova.scheduler.client.report [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.973992] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3da7b34d-669b-41c8-b0d9-9001e5749656 tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "ac9d308b-ac4b-4153-8e1d-7551213c8233" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.044s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.025554] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.385902] env[63197]: INFO nova.compute.manager [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Took 23.24 seconds to build instance. [ 586.397480] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Successfully created port: 7b119857-88a0-4493-9c78-b94cd950e5f3 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.429753] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.430471] env[63197]: ERROR nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Traceback (most recent call last): [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self.driver.spawn(context, instance, image_meta, [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] vm_ref = self.build_virtual_machine(instance, [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.430471] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] for vif in network_info: [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return self._sync_wrapper(fn, *args, **kwargs) [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self.wait() [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self[:] = self._gt.wait() [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return self._exit_event.wait() [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] result = hub.switch() [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.430860] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return self.greenlet.switch() [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] result = function(*args, **kwargs) [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] return func(*args, **kwargs) [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] raise e [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] nwinfo = self.network_api.allocate_for_instance( [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] created_port_ids = self._update_ports_for_instance( [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] with excutils.save_and_reraise_exception(): [ 586.431305] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] self.force_reraise() [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] raise self.value [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] updated_port = self._update_port( [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] _ensure_no_port_binding_failure(port) [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] raise exception.PortBindingFailed(port_id=port['id']) [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] nova.exception.PortBindingFailed: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. [ 586.431701] env[63197]: ERROR nova.compute.manager [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] [ 586.432928] env[63197]: DEBUG nova.compute.utils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.436430] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.239s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.438886] env[63197]: INFO nova.compute.claims [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.447288] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Build of instance 7ec7941d-183d-41d6-b2fb-6e795786c454 was re-scheduled: Binding failed for port 8569eb9e-ec1e-4cda-8885-daf63a6ef32e, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.447288] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.447288] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Acquiring lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.447288] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Acquired lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.447464] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.479778] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.517088] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 586.531952] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Releasing lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.531952] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.531952] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.531952] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ccf6d17e-9ce7-4840-9e29-5146b1c44d46 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.550190] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63818c0d-053e-44c1-a593-93cfef72b3f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.574641] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 586.575149] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 586.575149] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 586.578264] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 586.578613] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 586.578613] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 586.578846] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 586.579016] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 586.579248] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 586.579424] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 586.579604] env[63197]: DEBUG nova.virt.hardware [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.580875] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4a693c-de96-4ca3-968d-a2dd63bd346b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.590555] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 68ce48d7-e967-4cfc-a607-f6580bc14648 could not be found. [ 586.591881] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.591881] env[63197]: INFO nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Took 0.06 seconds to destroy the instance on the hypervisor. [ 586.591881] env[63197]: DEBUG oslo.service.loopingcall [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.591881] env[63197]: DEBUG nova.compute.manager [-] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.592097] env[63197]: DEBUG nova.network.neutron [-] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.597101] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef03a39b-874c-432b-8e13-9be91eef64fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.630899] env[63197]: DEBUG nova.network.neutron [-] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.817683] env[63197]: ERROR nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. [ 586.817683] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.817683] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.817683] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.817683] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.817683] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.817683] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.817683] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.817683] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.817683] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 586.817683] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.817683] env[63197]: ERROR nova.compute.manager raise self.value [ 586.817683] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.817683] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.817683] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.817683] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.818366] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.818366] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.818366] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. [ 586.818366] env[63197]: ERROR nova.compute.manager [ 586.818366] env[63197]: Traceback (most recent call last): [ 586.818366] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.818366] env[63197]: listener.cb(fileno) [ 586.818366] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.818366] env[63197]: result = function(*args, **kwargs) [ 586.818366] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.818366] env[63197]: return func(*args, **kwargs) [ 586.818366] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.818366] env[63197]: raise e [ 586.818366] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.818366] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 586.818366] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.818366] env[63197]: created_port_ids = self._update_ports_for_instance( [ 586.818366] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.818366] env[63197]: with excutils.save_and_reraise_exception(): [ 586.818366] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.818366] env[63197]: self.force_reraise() [ 586.818366] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.818366] env[63197]: raise self.value [ 586.818366] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.818366] env[63197]: updated_port = self._update_port( [ 586.818366] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.818366] env[63197]: _ensure_no_port_binding_failure(port) [ 586.818366] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.818366] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.819229] env[63197]: nova.exception.PortBindingFailed: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. [ 586.819229] env[63197]: Removing descriptor: 19 [ 586.819229] env[63197]: ERROR nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Traceback (most recent call last): [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] yield resources [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self.driver.spawn(context, instance, image_meta, [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.819229] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] vm_ref = self.build_virtual_machine(instance, [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] for vif in network_info: [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return self._sync_wrapper(fn, *args, **kwargs) [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self.wait() [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self[:] = self._gt.wait() [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return self._exit_event.wait() [ 586.819576] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] result = hub.switch() [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return self.greenlet.switch() [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] result = function(*args, **kwargs) [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return func(*args, **kwargs) [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] raise e [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] nwinfo = self.network_api.allocate_for_instance( [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.820099] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] created_port_ids = self._update_ports_for_instance( [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] with excutils.save_and_reraise_exception(): [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self.force_reraise() [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] raise self.value [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] updated_port = self._update_port( [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] _ensure_no_port_binding_failure(port) [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.820543] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] raise exception.PortBindingFailed(port_id=port['id']) [ 586.821743] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] nova.exception.PortBindingFailed: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. [ 586.821743] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] [ 586.821743] env[63197]: INFO nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Terminating instance [ 586.822409] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Acquiring lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.822800] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Acquired lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.823085] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.890471] env[63197]: DEBUG oslo_concurrency.lockutils [None req-27d93571-febb-4d1f-844e-824fdc54854c tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "99539d7f-cd93-4250-97cd-2449e03e517d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.751s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.890471] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "99539d7f-cd93-4250-97cd-2449e03e517d" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.835s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.890471] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] During sync_power_state the instance has a pending task (spawning). Skip. [ 586.890471] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "99539d7f-cd93-4250-97cd-2449e03e517d" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.973939] env[63197]: DEBUG nova.compute.manager [req-95525a17-6956-43a8-9c5d-6ce2a8885b12 req-41ab0b22-4c28-4832-9649-b8dc59bfec0e service nova] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Received event network-vif-deleted-56564457-bbf2-48f2-af44-708898339d9f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.987393] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.003155] env[63197]: DEBUG nova.compute.manager [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Received event network-changed-e312e7b2-5270-4cf6-8678-7a1af450ad88 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.003155] env[63197]: DEBUG nova.compute.manager [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Refreshing instance network info cache due to event network-changed-e312e7b2-5270-4cf6-8678-7a1af450ad88. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 587.003155] env[63197]: DEBUG oslo_concurrency.lockutils [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] Acquiring lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.004206] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.133407] env[63197]: DEBUG nova.network.neutron [-] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.218940] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.226541] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Acquiring lock "b52f3d4b-7699-484d-8b4e-84393848fb37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.226987] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Lock "b52f3d4b-7699-484d-8b4e-84393848fb37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.360383] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.391747] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.538723] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.642306] env[63197]: INFO nova.compute.manager [-] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Took 1.05 seconds to deallocate network for instance. [ 587.648150] env[63197]: DEBUG nova.compute.claims [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.648150] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.724172] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Releasing lock "refresh_cache-7ec7941d-183d-41d6-b2fb-6e795786c454" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.724172] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.724625] env[63197]: DEBUG nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.726120] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.792339] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.894108] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ab2660-91f3-4181-ba10-0bda496a0cff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.912399] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929b32b5-b249-41fe-9fbc-d67f73942337 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.931326] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.964938] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd23019-c61f-41f2-98d7-d736085ed2a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.973270] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb4e695-6a05-43e9-98a0-52ee965c757c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.989779] env[63197]: DEBUG nova.compute.provider_tree [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.039655] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Releasing lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.040185] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 588.040396] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 588.040697] env[63197]: DEBUG oslo_concurrency.lockutils [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] Acquired lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.040856] env[63197]: DEBUG nova.network.neutron [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Refreshing network info cache for port e312e7b2-5270-4cf6-8678-7a1af450ad88 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.043183] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e2c31ad-d8db-4b0b-a735-1450aedacc2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.055601] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-115765b9-4564-4f8b-807e-f8e0edda44d2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.083708] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37280bd4-3cd8-4910-97de-0cc7494abc22 could not be found. [ 588.083896] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 588.084085] env[63197]: INFO nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Took 0.04 seconds to destroy the instance on the hypervisor. [ 588.084323] env[63197]: DEBUG oslo.service.loopingcall [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.084545] env[63197]: DEBUG nova.compute.manager [-] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.084633] env[63197]: DEBUG nova.network.neutron [-] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.126275] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Acquiring lock "a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.127442] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Lock "a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.140521] env[63197]: DEBUG nova.network.neutron [-] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.295068] env[63197]: DEBUG nova.network.neutron [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.435790] env[63197]: INFO nova.compute.manager [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Rebuilding instance [ 588.490851] env[63197]: DEBUG nova.compute.manager [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 588.491534] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aacf26b0-3c1e-497a-91fe-122ebfd465a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.497426] env[63197]: DEBUG nova.scheduler.client.report [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.610041] env[63197]: DEBUG nova.network.neutron [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.643201] env[63197]: DEBUG nova.network.neutron [-] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.782386] env[63197]: DEBUG nova.network.neutron [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.798911] env[63197]: INFO nova.compute.manager [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] Took 1.07 seconds to deallocate network for instance. [ 589.009407] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.011015] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.018114] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 589.018781] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.012s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.023198] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f80da5e-e7b2-4cae-97b1-1094f8198146 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.032321] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 589.032321] env[63197]: value = "task-1364028" [ 589.032321] env[63197]: _type = "Task" [ 589.032321] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.049412] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364028, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.146746] env[63197]: INFO nova.compute.manager [-] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Took 1.06 seconds to deallocate network for instance. [ 589.150646] env[63197]: DEBUG nova.compute.claims [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 589.150908] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.219123] env[63197]: ERROR nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. [ 589.219123] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.219123] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.219123] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.219123] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.219123] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.219123] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.219123] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.219123] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.219123] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 589.219123] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.219123] env[63197]: ERROR nova.compute.manager raise self.value [ 589.219123] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.219123] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.219123] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.219123] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.219650] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.219650] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.219650] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. [ 589.219650] env[63197]: ERROR nova.compute.manager [ 589.219650] env[63197]: Traceback (most recent call last): [ 589.219650] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.219650] env[63197]: listener.cb(fileno) [ 589.219650] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.219650] env[63197]: result = function(*args, **kwargs) [ 589.219650] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.219650] env[63197]: return func(*args, **kwargs) [ 589.219650] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.219650] env[63197]: raise e [ 589.219650] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.219650] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 589.219650] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.219650] env[63197]: created_port_ids = self._update_ports_for_instance( [ 589.219650] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.219650] env[63197]: with excutils.save_and_reraise_exception(): [ 589.219650] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.219650] env[63197]: self.force_reraise() [ 589.219650] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.219650] env[63197]: raise self.value [ 589.219650] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.219650] env[63197]: updated_port = self._update_port( [ 589.219650] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.219650] env[63197]: _ensure_no_port_binding_failure(port) [ 589.219650] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.219650] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.220609] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. [ 589.220609] env[63197]: Removing descriptor: 18 [ 589.220609] env[63197]: ERROR nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Traceback (most recent call last): [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] yield resources [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self.driver.spawn(context, instance, image_meta, [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.220609] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] vm_ref = self.build_virtual_machine(instance, [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] for vif in network_info: [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return self._sync_wrapper(fn, *args, **kwargs) [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self.wait() [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self[:] = self._gt.wait() [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return self._exit_event.wait() [ 589.221375] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] result = hub.switch() [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return self.greenlet.switch() [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] result = function(*args, **kwargs) [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return func(*args, **kwargs) [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] raise e [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] nwinfo = self.network_api.allocate_for_instance( [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.221779] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] created_port_ids = self._update_ports_for_instance( [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] with excutils.save_and_reraise_exception(): [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self.force_reraise() [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] raise self.value [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] updated_port = self._update_port( [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] _ensure_no_port_binding_failure(port) [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.222186] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] raise exception.PortBindingFailed(port_id=port['id']) [ 589.222516] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] nova.exception.PortBindingFailed: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. [ 589.222516] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] [ 589.222516] env[63197]: INFO nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Terminating instance [ 589.222516] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.222643] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquired lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.223322] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.286147] env[63197]: DEBUG oslo_concurrency.lockutils [req-cef67866-cf03-490c-9d6c-e17ee3d33bd7 req-65e66fd6-dc2b-445c-9a47-699b1957c78c service nova] Releasing lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.523131] env[63197]: DEBUG nova.compute.utils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 589.524499] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 589.524659] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 589.552815] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364028, 'name': PowerOffVM_Task, 'duration_secs': 0.175716} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.553095] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 589.553305] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.554321] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb68d91-4bfb-4757-9ab7-3f76d14fe71d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.567100] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 589.570801] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c90d149-08ae-4846-a104-d6a0d1f65ace {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.599140] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 589.599392] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 589.602097] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleting the datastore file [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 589.602097] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-651b4a54-7aa2-4fc2-b6b3-a4721210500e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.608334] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 589.608334] env[63197]: value = "task-1364030" [ 589.608334] env[63197]: _type = "Task" [ 589.608334] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.617683] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.677984] env[63197]: DEBUG nova.policy [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb0b157fe55f452ba29d9142a058b1c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1a5192271c84747804e24225b6df8d4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.747490] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.773136] env[63197]: DEBUG nova.compute.manager [None req-9698ed7b-2a18-453a-9461-22cf32a559b1 tempest-ServerDiagnosticsV248Test-400240072 tempest-ServerDiagnosticsV248Test-400240072-project-admin] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 589.774319] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8332390-5e29-4a13-bc31-ee398ecc7caf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.783037] env[63197]: INFO nova.compute.manager [None req-9698ed7b-2a18-453a-9461-22cf32a559b1 tempest-ServerDiagnosticsV248Test-400240072 tempest-ServerDiagnosticsV248Test-400240072-project-admin] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Retrieving diagnostics [ 589.783331] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3f335f-addf-4b6a-9fa2-0620b52002d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.848517] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.865511] env[63197]: INFO nova.scheduler.client.report [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Deleted allocations for instance 7ec7941d-183d-41d6-b2fb-6e795786c454 [ 590.020509] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538ba851-7d10-4dc4-9646-da8bd01b6911 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.031891] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7710a1b-c1ee-4c5e-9690-c409abb2ac56 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.034947] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.070123] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b8eb4d-ec46-4136-8aa0-b0780d1e7a0e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.077022] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f109238d-5e68-4c49-9c1d-4b89a93ec25c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.090694] env[63197]: DEBUG nova.compute.provider_tree [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.117814] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103024} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.118078] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 590.118887] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 590.118887] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.132966] env[63197]: DEBUG nova.compute.manager [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Received event network-changed-7b119857-88a0-4493-9c78-b94cd950e5f3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.133485] env[63197]: DEBUG nova.compute.manager [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Refreshing instance network info cache due to event network-changed-7b119857-88a0-4493-9c78-b94cd950e5f3. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 590.133901] env[63197]: DEBUG oslo_concurrency.lockutils [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] Acquiring lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.205462] env[63197]: DEBUG nova.compute.manager [req-a542b1f8-d487-40fb-b733-cb54fba0ec2e req-3da76a6b-ebd5-410b-8205-d85df1b8c238 service nova] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Received event network-vif-deleted-e312e7b2-5270-4cf6-8678-7a1af450ad88 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.355353] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Releasing lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.355876] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.358022] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.358022] env[63197]: DEBUG oslo_concurrency.lockutils [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] Acquired lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.358022] env[63197]: DEBUG nova.network.neutron [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Refreshing network info cache for port 7b119857-88a0-4493-9c78-b94cd950e5f3 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 590.358022] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82b9b28e-bad4-4d19-a0df-55975a880892 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.373563] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7208083b-58c3-4ef4-8d5b-1991b72c2525 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.388982] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58f079f5-2747-4012-8fe9-45c241e4f526 tempest-AttachInterfacesUnderV243Test-1048207652 tempest-AttachInterfacesUnderV243Test-1048207652-project-member] Lock "7ec7941d-183d-41d6-b2fb-6e795786c454" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.453s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.390773] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "7ec7941d-183d-41d6-b2fb-6e795786c454" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.338s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.394429] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 7ec7941d-183d-41d6-b2fb-6e795786c454] During sync_power_state the instance has a pending task (spawning). Skip. [ 590.394750] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "7ec7941d-183d-41d6-b2fb-6e795786c454" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.004s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.410741] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a8315374-abb6-4663-bdb5-c4d88cdbbaa4 could not be found. [ 590.410978] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.411178] env[63197]: INFO nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Took 0.06 seconds to destroy the instance on the hypervisor. [ 590.411423] env[63197]: DEBUG oslo.service.loopingcall [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.411881] env[63197]: DEBUG nova.compute.manager [-] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.411974] env[63197]: DEBUG nova.network.neutron [-] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 590.437627] env[63197]: DEBUG nova.network.neutron [-] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.440819] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Successfully created port: 8f7b4666-c370-4d86-85da-b99b1dcbc6db {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.593485] env[63197]: DEBUG nova.scheduler.client.report [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.894217] env[63197]: DEBUG nova.network.neutron [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.898435] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 590.943639] env[63197]: DEBUG nova.network.neutron [-] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.021519] env[63197]: DEBUG nova.network.neutron [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.046527] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.079296] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.079537] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.079696] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.079869] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.080075] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.080309] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.080625] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.080822] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.081036] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.081243] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.081477] env[63197]: DEBUG nova.virt.hardware [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.082485] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a0d13c-2be1-447a-a74f-c185d51b2f22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.090690] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44173768-5684-499a-89d6-276f71aa84c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.106158] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.087s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.106915] env[63197]: ERROR nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Traceback (most recent call last): [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self.driver.spawn(context, instance, image_meta, [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] vm_ref = self.build_virtual_machine(instance, [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.106915] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] for vif in network_info: [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return self._sync_wrapper(fn, *args, **kwargs) [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self.wait() [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self[:] = self._gt.wait() [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return self._exit_event.wait() [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] result = hub.switch() [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.107306] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return self.greenlet.switch() [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] result = function(*args, **kwargs) [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] return func(*args, **kwargs) [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] raise e [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] nwinfo = self.network_api.allocate_for_instance( [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] created_port_ids = self._update_ports_for_instance( [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] with excutils.save_and_reraise_exception(): [ 591.107837] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] self.force_reraise() [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] raise self.value [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] updated_port = self._update_port( [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] _ensure_no_port_binding_failure(port) [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] raise exception.PortBindingFailed(port_id=port['id']) [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] nova.exception.PortBindingFailed: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. [ 591.108238] env[63197]: ERROR nova.compute.manager [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] [ 591.108652] env[63197]: DEBUG nova.compute.utils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.110468] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Build of instance a892c650-d8e8-4eaa-b025-cf330478aabf was re-scheduled: Binding failed for port 9d17dab0-f776-4eb7-8729-0ede74a50ddd, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 591.110468] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 591.110468] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Acquiring lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.110666] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Acquired lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.110799] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.111762] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.462s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.153064] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.153396] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.153693] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.153908] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.154111] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.154297] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.154538] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.154736] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.154943] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.155176] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.155407] env[63197]: DEBUG nova.virt.hardware [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.156316] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cde7d9b-3701-4811-ae35-ec7812aba6f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.166896] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ad9cb5-4207-48a3-a017-c1ff9105f5cb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.179449] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 591.186724] env[63197]: DEBUG oslo.service.loopingcall [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.186724] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 591.186724] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d277883-c4df-4beb-a9b4-204f5b1c3bb2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.211230] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 591.211230] env[63197]: value = "task-1364031" [ 591.211230] env[63197]: _type = "Task" [ 591.211230] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.219585] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364031, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.431867] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.445998] env[63197]: INFO nova.compute.manager [-] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Took 1.03 seconds to deallocate network for instance. [ 591.453668] env[63197]: DEBUG nova.compute.claims [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.453873] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.523617] env[63197]: DEBUG oslo_concurrency.lockutils [req-3a9842ca-df7a-4991-a86c-96357342f338 req-62502c89-59ad-47b4-8b1f-5b5d61954fb0 service nova] Releasing lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.648439] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.708819] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.710310] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.710615] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.710900] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.711086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.716822] env[63197]: INFO nova.compute.manager [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Terminating instance [ 591.718916] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "refresh_cache-9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.719122] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquired lock "refresh_cache-9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.719321] env[63197]: DEBUG nova.network.neutron [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.734131] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364031, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.854711] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.958680] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb96ccd-bc9d-4bc0-b813-7c7ff0a5b36a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.968076] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cc2439-978c-418b-ba2a-b0300a0c8bae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.007309] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2365e0f-4d30-4695-8300-be9a5ee576d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.019680] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67648467-8da4-459c-906f-cba60609a78d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.034788] env[63197]: DEBUG nova.compute.provider_tree [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.225688] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364031, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.270664] env[63197]: DEBUG nova.network.neutron [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.358221] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Releasing lock "refresh_cache-a892c650-d8e8-4eaa-b025-cf330478aabf" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.358507] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.358649] env[63197]: DEBUG nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.358813] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.364148] env[63197]: DEBUG nova.network.neutron [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.395922] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.538153] env[63197]: DEBUG nova.scheduler.client.report [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.614131] env[63197]: ERROR nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. [ 592.614131] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.614131] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.614131] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.614131] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.614131] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.614131] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.614131] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.614131] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.614131] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 592.614131] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.614131] env[63197]: ERROR nova.compute.manager raise self.value [ 592.614131] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.614131] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.614131] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.614131] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.614715] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.614715] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.614715] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. [ 592.614715] env[63197]: ERROR nova.compute.manager [ 592.614715] env[63197]: Traceback (most recent call last): [ 592.614715] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.614715] env[63197]: listener.cb(fileno) [ 592.614715] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.614715] env[63197]: result = function(*args, **kwargs) [ 592.614715] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.614715] env[63197]: return func(*args, **kwargs) [ 592.614715] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.614715] env[63197]: raise e [ 592.614715] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.614715] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 592.614715] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.614715] env[63197]: created_port_ids = self._update_ports_for_instance( [ 592.614715] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.614715] env[63197]: with excutils.save_and_reraise_exception(): [ 592.614715] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.614715] env[63197]: self.force_reraise() [ 592.614715] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.614715] env[63197]: raise self.value [ 592.614715] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.614715] env[63197]: updated_port = self._update_port( [ 592.614715] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.614715] env[63197]: _ensure_no_port_binding_failure(port) [ 592.614715] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.614715] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.615670] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. [ 592.615670] env[63197]: Removing descriptor: 18 [ 592.615670] env[63197]: ERROR nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Traceback (most recent call last): [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] yield resources [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self.driver.spawn(context, instance, image_meta, [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.615670] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] vm_ref = self.build_virtual_machine(instance, [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] for vif in network_info: [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return self._sync_wrapper(fn, *args, **kwargs) [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self.wait() [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self[:] = self._gt.wait() [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return self._exit_event.wait() [ 592.616110] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] result = hub.switch() [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return self.greenlet.switch() [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] result = function(*args, **kwargs) [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return func(*args, **kwargs) [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] raise e [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] nwinfo = self.network_api.allocate_for_instance( [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.616470] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] created_port_ids = self._update_ports_for_instance( [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] with excutils.save_and_reraise_exception(): [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self.force_reraise() [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] raise self.value [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] updated_port = self._update_port( [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] _ensure_no_port_binding_failure(port) [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.616864] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] raise exception.PortBindingFailed(port_id=port['id']) [ 592.617197] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] nova.exception.PortBindingFailed: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. [ 592.617197] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] [ 592.617197] env[63197]: INFO nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Terminating instance [ 592.617197] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Acquiring lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.617197] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Acquired lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.617197] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.728637] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Acquiring lock "4f2d4315-d533-431a-970a-ea655febf0ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.728637] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Lock "4f2d4315-d533-431a-970a-ea655febf0ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.733810] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364031, 'name': CreateVM_Task, 'duration_secs': 1.279148} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.735180] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 592.735180] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.735180] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.735180] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 592.739018] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc54e1c0-d8fa-43fc-8943-f306f6a688ba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.747233] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 592.747233] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522c3f34-0bd2-ad8e-ea44-d3ca71eae082" [ 592.747233] env[63197]: _type = "Task" [ 592.747233] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.757724] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522c3f34-0bd2-ad8e-ea44-d3ca71eae082, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.763675] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "47da4c8f-a4cb-4d7f-81c0-7c3019169b56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.763891] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "47da4c8f-a4cb-4d7f-81c0-7c3019169b56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.868787] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Releasing lock "refresh_cache-9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.868787] env[63197]: DEBUG nova.compute.manager [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 592.868787] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 592.868787] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afae6282-4365-4007-beea-d3e0977a86b3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.876722] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 592.877039] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bd82d8f-d85a-4614-835d-e6b9d3418039 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.882815] env[63197]: DEBUG oslo_vmware.api [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 592.882815] env[63197]: value = "task-1364032" [ 592.882815] env[63197]: _type = "Task" [ 592.882815] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.891639] env[63197]: DEBUG oslo_vmware.api [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364032, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.898825] env[63197]: DEBUG nova.network.neutron [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.044800] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.933s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.045448] env[63197]: ERROR nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Traceback (most recent call last): [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self.driver.spawn(context, instance, image_meta, [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] vm_ref = self.build_virtual_machine(instance, [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 593.045448] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] for vif in network_info: [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return self._sync_wrapper(fn, *args, **kwargs) [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self.wait() [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self[:] = self._gt.wait() [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return self._exit_event.wait() [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] result = hub.switch() [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 593.045892] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return self.greenlet.switch() [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] result = function(*args, **kwargs) [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] return func(*args, **kwargs) [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] raise e [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] nwinfo = self.network_api.allocate_for_instance( [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] created_port_ids = self._update_ports_for_instance( [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] with excutils.save_and_reraise_exception(): [ 593.046637] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] self.force_reraise() [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] raise self.value [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] updated_port = self._update_port( [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] _ensure_no_port_binding_failure(port) [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] raise exception.PortBindingFailed(port_id=port['id']) [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] nova.exception.PortBindingFailed: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. [ 593.047247] env[63197]: ERROR nova.compute.manager [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] [ 593.047692] env[63197]: DEBUG nova.compute.utils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 593.047692] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.218s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.050214] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Build of instance c0742ebb-c176-44cf-ae13-d778cc6364cb was re-scheduled: Binding failed for port 0042be2f-088c-4b82-a376-3dfabacf34fd, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 593.050654] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 593.050876] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Acquiring lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.051032] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Acquired lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.051197] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.152436] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.267729] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522c3f34-0bd2-ad8e-ea44-d3ca71eae082, 'name': SearchDatastore_Task, 'duration_secs': 0.017051} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.271810] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.272065] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 593.272370] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.273309] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.273309] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 593.275448] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d756cc95-7200-4edf-aa4b-a14c8b497215 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.278452] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Acquiring lock "6abf52eb-a394-477f-95e6-87cddecd4bd0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.278771] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Lock "6abf52eb-a394-477f-95e6-87cddecd4bd0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.285140] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 593.285323] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 593.286049] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b82f339-ebce-48c4-b66d-49f1f0e7428b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.295027] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 593.295027] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5297ee7f-e9f6-dd68-760a-cb9fe9bbae39" [ 593.295027] env[63197]: _type = "Task" [ 593.295027] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.308381] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5297ee7f-e9f6-dd68-760a-cb9fe9bbae39, 'name': SearchDatastore_Task, 'duration_secs': 0.008176} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.308666] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb4cc1b1-9b4c-41c4-815d-5f5f8075eec3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.323185] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 593.323185] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bb0cc0-ef23-c6da-c08d-8ea4fb41b0d4" [ 593.323185] env[63197]: _type = "Task" [ 593.323185] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.335039] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bb0cc0-ef23-c6da-c08d-8ea4fb41b0d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.354166] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.393421] env[63197]: DEBUG oslo_vmware.api [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364032, 'name': PowerOffVM_Task, 'duration_secs': 0.106938} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.393710] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 593.396291] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 593.399028] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67cd3711-c66d-40b3-b23b-92df5b7799bf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.401503] env[63197]: INFO nova.compute.manager [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] Took 1.04 seconds to deallocate network for instance. [ 593.423101] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 593.423101] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 593.423101] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Deleting the datastore file [datastore2] 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 593.423101] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10035e8f-2006-450f-a7d1-67e36681043d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.430731] env[63197]: DEBUG oslo_vmware.api [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for the task: (returnval){ [ 593.430731] env[63197]: value = "task-1364034" [ 593.430731] env[63197]: _type = "Task" [ 593.430731] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.443433] env[63197]: DEBUG oslo_vmware.api [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364034, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.602198] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.646174] env[63197]: DEBUG nova.compute.manager [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Received event network-vif-deleted-7b119857-88a0-4493-9c78-b94cd950e5f3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.646374] env[63197]: DEBUG nova.compute.manager [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Received event network-changed-8f7b4666-c370-4d86-85da-b99b1dcbc6db {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.646539] env[63197]: DEBUG nova.compute.manager [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Refreshing instance network info cache due to event network-changed-8f7b4666-c370-4d86-85da-b99b1dcbc6db. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 593.649607] env[63197]: DEBUG oslo_concurrency.lockutils [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] Acquiring lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.773190] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.811175] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Acquiring lock "0d7e1031-896b-4ec3-9da0-74d65b1ef78a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.811175] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Lock "0d7e1031-896b-4ec3-9da0-74d65b1ef78a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.832843] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bb0cc0-ef23-c6da-c08d-8ea4fb41b0d4, 'name': SearchDatastore_Task, 'duration_secs': 0.025429} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.833255] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.833737] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 593.833990] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5847d9ea-7a03-49b2-ac8d-5cd626912069 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.845921] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 593.845921] env[63197]: value = "task-1364035" [ 593.845921] env[63197]: _type = "Task" [ 593.845921] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.856968] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Releasing lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.857428] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.857690] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.857882] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364035, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.858123] env[63197]: DEBUG oslo_concurrency.lockutils [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] Acquired lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.858289] env[63197]: DEBUG nova.network.neutron [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Refreshing network info cache for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 593.859340] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06b6148c-0d52-4a9c-87c3-41c2e6861111 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.874160] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d710a630-f319-46c2-8a10-e88fd8653288 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.896986] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06de0427-5e1a-469c-ae45-9dc72336d3cc could not be found. [ 593.897682] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.897682] env[63197]: INFO nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.897682] env[63197]: DEBUG oslo.service.loopingcall [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.900558] env[63197]: DEBUG nova.compute.manager [-] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.900655] env[63197]: DEBUG nova.network.neutron [-] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.925675] env[63197]: DEBUG nova.network.neutron [-] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.946480] env[63197]: DEBUG oslo_vmware.api [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Task: {'id': task-1364034, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097777} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.946835] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 593.946942] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 593.947137] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.947329] env[63197]: INFO nova.compute.manager [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Took 1.08 seconds to destroy the instance on the hypervisor. [ 593.947560] env[63197]: DEBUG oslo.service.loopingcall [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.947959] env[63197]: DEBUG nova.compute.manager [-] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.948074] env[63197]: DEBUG nova.network.neutron [-] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.952705] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefd0e7e-b46f-4f54-840b-05fd45bcd469 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.962041] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d85020-2991-4c86-a6c5-11678a28dea6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.002036] env[63197]: DEBUG nova.network.neutron [-] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.003156] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a47a3ed-d403-42e4-9451-d4a0ebb41572 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.011223] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fbfce3-0a3d-4056-bb2d-2863b26b080d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.027432] env[63197]: DEBUG nova.compute.provider_tree [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.277898] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Releasing lock "refresh_cache-c0742ebb-c176-44cf-ae13-d778cc6364cb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.277898] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 594.277898] env[63197]: DEBUG nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.277898] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.309931] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.360959] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364035, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.428174] env[63197]: DEBUG nova.network.neutron [-] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.458676] env[63197]: INFO nova.scheduler.client.report [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Deleted allocations for instance a892c650-d8e8-4eaa-b025-cf330478aabf [ 594.507133] env[63197]: DEBUG nova.network.neutron [-] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.530179] env[63197]: DEBUG nova.network.neutron [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.533919] env[63197]: DEBUG nova.scheduler.client.report [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.670252] env[63197]: DEBUG nova.network.neutron [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.802022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "11c6fcd4-87d3-4ceb-98eb-a46ca2910584" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.802022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "11c6fcd4-87d3-4ceb-98eb-a46ca2910584" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.815509] env[63197]: DEBUG nova.network.neutron [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.862427] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364035, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548981} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.862803] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 594.863072] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 594.863406] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-070e2ffd-9487-4a65-b5e1-207e607e4d77 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.871626] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 594.871626] env[63197]: value = "task-1364036" [ 594.871626] env[63197]: _type = "Task" [ 594.871626] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.881090] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364036, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.934098] env[63197]: INFO nova.compute.manager [-] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Took 1.03 seconds to deallocate network for instance. [ 594.935847] env[63197]: DEBUG nova.compute.claims [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.936063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.967391] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4218c637-baac-4a79-bd5c-5fb3361fa010 tempest-ServerAddressesTestJSON-1832323071 tempest-ServerAddressesTestJSON-1832323071-project-member] Lock "a892c650-d8e8-4eaa-b025-cf330478aabf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.948s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.968693] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "a892c650-d8e8-4eaa-b025-cf330478aabf" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 11.915s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.968883] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a892c650-d8e8-4eaa-b025-cf330478aabf] During sync_power_state the instance has a pending task (spawning). Skip. [ 594.969063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "a892c650-d8e8-4eaa-b025-cf330478aabf" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.014585] env[63197]: INFO nova.compute.manager [-] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Took 1.07 seconds to deallocate network for instance. [ 595.041789] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.994s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.042323] env[63197]: ERROR nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Traceback (most recent call last): [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self.driver.spawn(context, instance, image_meta, [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] vm_ref = self.build_virtual_machine(instance, [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.042323] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] for vif in network_info: [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return self._sync_wrapper(fn, *args, **kwargs) [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self.wait() [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self[:] = self._gt.wait() [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return self._exit_event.wait() [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] result = hub.switch() [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.042947] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return self.greenlet.switch() [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] result = function(*args, **kwargs) [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] return func(*args, **kwargs) [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] raise e [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] nwinfo = self.network_api.allocate_for_instance( [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] created_port_ids = self._update_ports_for_instance( [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] with excutils.save_and_reraise_exception(): [ 595.043347] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] self.force_reraise() [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] raise self.value [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] updated_port = self._update_port( [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] _ensure_no_port_binding_failure(port) [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] raise exception.PortBindingFailed(port_id=port['id']) [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] nova.exception.PortBindingFailed: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. [ 595.043881] env[63197]: ERROR nova.compute.manager [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] [ 595.045541] env[63197]: DEBUG nova.compute.utils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 595.045541] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.475s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.045541] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.045541] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 595.045541] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.041s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.046889] env[63197]: INFO nova.compute.claims [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.050148] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Build of instance 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e was re-scheduled: Binding failed for port 9cd1d79f-00a0-4402-b99c-6f48616e1579, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 595.050341] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 595.050594] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.050744] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.055174] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 595.058096] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58025e85-9666-4d1e-9318-a42dabf36298 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.067149] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de80e1b-b2ef-4ab6-a0ed-979f32575390 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.090037] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf84254c-c9b9-4dc8-8e9c-6858a2d46e36 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.095503] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688df5c6-9749-4282-9c57-9931ea26bafd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.132728] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181495MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 595.132898] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.174255] env[63197]: DEBUG oslo_concurrency.lockutils [req-6c961cac-e617-4f90-ad3c-1b6f02886ae2 req-1d26a4cc-9ca1-4500-a234-1c7942d54658 service nova] Releasing lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.318905] env[63197]: INFO nova.compute.manager [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] Took 1.04 seconds to deallocate network for instance. [ 595.385917] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364036, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071375} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.390368] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 595.391860] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d9f943-f8ae-4d07-9256-f263b3dca20d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.412086] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 595.412708] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dba4188f-b54f-449d-8738-4083c7365fcf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.435019] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 595.435019] env[63197]: value = "task-1364037" [ 595.435019] env[63197]: _type = "Task" [ 595.435019] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.444784] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364037, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.469979] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.528729] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.615842] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.798434] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.945448] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364037, 'name': ReconfigVM_Task, 'duration_secs': 0.278657} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.945799] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Reconfigured VM instance instance-00000007 to attach disk [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 595.946458] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be3d31f5-4cb5-4674-80e0-9d334d590861 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.954292] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 595.954292] env[63197]: value = "task-1364038" [ 595.954292] env[63197]: _type = "Task" [ 595.954292] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.971337] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364038, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.001973] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.302418] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "refresh_cache-01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.302418] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 596.302418] env[63197]: DEBUG nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.302665] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 596.351311] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.364610] env[63197]: INFO nova.scheduler.client.report [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Deleted allocations for instance c0742ebb-c176-44cf-ae13-d778cc6364cb [ 596.383178] env[63197]: DEBUG nova.compute.manager [req-6756173d-58a0-4c9e-9973-bc7ef5980be9 req-08f32878-b0e9-4e5a-9c68-4d1ea953e1c3 service nova] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Received event network-vif-deleted-8f7b4666-c370-4d86-85da-b99b1dcbc6db {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.471555] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364038, 'name': Rename_Task, 'duration_secs': 0.123789} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.472779] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 596.475342] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19f9de6-29f3-4222-bea5-3afc06c7f5e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.475996] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61b486e3-b59f-4ac4-8c3a-c49d0ab8a086 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.482066] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d54eb0-dff2-45ab-a209-eead3bdd3c9b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.487322] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 596.487322] env[63197]: value = "task-1364039" [ 596.487322] env[63197]: _type = "Task" [ 596.487322] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.529368] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bbd990e-e2b9-4a8f-bb48-616eeb251668 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.536761] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364039, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.541805] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6336d9ed-3fce-4c98-9eb5-074385d1c97d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.562552] env[63197]: DEBUG nova.compute.provider_tree [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.862663] env[63197]: DEBUG nova.network.neutron [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.877711] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0114ba3c-dda9-42dd-b8f3-62f4593f8e74 tempest-ServerDiagnosticsTest-1745331318 tempest-ServerDiagnosticsTest-1745331318-project-member] Lock "c0742ebb-c176-44cf-ae13-d778cc6364cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.472s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.879134] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "c0742ebb-c176-44cf-ae13-d778cc6364cb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 13.824s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.879383] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: c0742ebb-c176-44cf-ae13-d778cc6364cb] During sync_power_state the instance has a pending task (spawning). Skip. [ 596.879622] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "c0742ebb-c176-44cf-ae13-d778cc6364cb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.999671] env[63197]: DEBUG oslo_vmware.api [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364039, 'name': PowerOnVM_Task, 'duration_secs': 0.487837} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.000178] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 597.000332] env[63197]: DEBUG nova.compute.manager [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 597.001084] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3002fe0-590e-4770-8ae7-e7d97f9ab76f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.066383] env[63197]: DEBUG nova.scheduler.client.report [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.368873] env[63197]: INFO nova.compute.manager [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] Took 1.07 seconds to deallocate network for instance. [ 597.381528] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.519454] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.571511] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.572060] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.575248] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.928s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.925405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.086500] env[63197]: DEBUG nova.compute.utils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.092140] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.092318] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 598.197488] env[63197]: DEBUG nova.policy [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3424c6474e0f4cfd9ad8d11116a8d049', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f425c830852a4e0e96f3e0ee1437a30d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.425046] env[63197]: INFO nova.scheduler.client.report [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Deleted allocations for instance 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e [ 598.471672] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f34133-e1bd-4688-b794-c3402eb89429 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.486402] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c61256c-c012-4a88-b774-cdd243574f26 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.524988] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127dd8d1-3da5-4a72-b902-4162120460cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.532489] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88047d94-3ef6-43ba-b798-969ed8900a79 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.546841] env[63197]: DEBUG nova.compute.provider_tree [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.552449] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Acquiring lock "68b27a14-f52b-4d89-9cdc-be0529428eb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.552846] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Lock "68b27a14-f52b-4d89-9cdc-be0529428eb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.593423] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 598.940120] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490e8455-059f-4963-aed8-e09bed48338c tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.124s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.941577] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 15.888s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.941772] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e] During sync_power_state the instance has a pending task (spawning). Skip. [ 598.941934] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "01af6a6f-55e1-4ebb-bbc5-c4588bad0d9e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.977440] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Successfully created port: af208dde-7e08-4ea6-98b0-79296d2da0ae {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.053915] env[63197]: DEBUG nova.scheduler.client.report [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.173415] env[63197]: INFO nova.compute.manager [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Rebuilding instance [ 599.237883] env[63197]: DEBUG nova.compute.manager [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 599.238791] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27857ef-262a-4221-bffc-3d00c7e2e6cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.445259] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 599.560800] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.986s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.561622] env[63197]: ERROR nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Traceback (most recent call last): [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self.driver.spawn(context, instance, image_meta, [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] vm_ref = self.build_virtual_machine(instance, [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.561622] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] for vif in network_info: [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return self._sync_wrapper(fn, *args, **kwargs) [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self.wait() [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self[:] = self._gt.wait() [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return self._exit_event.wait() [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] result = hub.switch() [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.561960] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return self.greenlet.switch() [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] result = function(*args, **kwargs) [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] return func(*args, **kwargs) [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] raise e [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] nwinfo = self.network_api.allocate_for_instance( [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] created_port_ids = self._update_ports_for_instance( [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] with excutils.save_and_reraise_exception(): [ 599.562404] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] self.force_reraise() [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] raise self.value [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] updated_port = self._update_port( [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] _ensure_no_port_binding_failure(port) [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] raise exception.PortBindingFailed(port_id=port['id']) [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] nova.exception.PortBindingFailed: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. [ 599.562753] env[63197]: ERROR nova.compute.manager [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] [ 599.566420] env[63197]: DEBUG nova.compute.utils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.566420] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.633s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.566420] env[63197]: INFO nova.compute.claims [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.570172] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Build of instance 68ce48d7-e967-4cfc-a607-f6580bc14648 was re-scheduled: Binding failed for port 56564457-bbf2-48f2-af44-708898339d9f, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.571144] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.572047] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Acquiring lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.572378] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Acquired lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.572666] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 599.607554] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 599.640254] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 599.640254] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 599.640254] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.640566] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 599.640566] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.640566] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 599.640794] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 599.641382] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 599.641712] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 599.642409] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 599.642740] env[63197]: DEBUG nova.virt.hardware [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 599.643788] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8f8f95-7812-41ef-b1f7-1d81be373c76 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.659474] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89aa4971-931c-4f66-8f17-ea81c36753e2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.753312] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 599.753312] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-614876ad-132e-4a6f-8bb8-54969dd1156a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.763147] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 599.763147] env[63197]: value = "task-1364040" [ 599.763147] env[63197]: _type = "Task" [ 599.763147] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.777956] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364040, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.984499] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.123882] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.260867] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.277800] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364040, 'name': PowerOffVM_Task, 'duration_secs': 0.184007} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.279401] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 600.279631] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 600.280488] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997430e3-eecf-402e-b4be-b4cf68446df5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.288503] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 600.288630] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a757d47a-a595-4876-8812-edc15474e7f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.313607] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 600.313607] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 600.313607] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Deleting the datastore file [datastore1] 99539d7f-cd93-4250-97cd-2449e03e517d {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 600.313607] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c0acfd9-3174-4ca1-87a3-52d1985c8f15 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.320149] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 600.320149] env[63197]: value = "task-1364042" [ 600.320149] env[63197]: _type = "Task" [ 600.320149] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.328344] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364042, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.763640] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Releasing lock "refresh_cache-68ce48d7-e967-4cfc-a607-f6580bc14648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.763915] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 600.764049] env[63197]: DEBUG nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.764452] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 600.802770] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.832339] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364042, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091117} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.832582] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 600.832759] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 600.832929] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 600.991131] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70396d9-1dfe-4162-8dd0-05bc37ce0641 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.999402] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61027ad1-d445-418c-b7ff-d91c419256a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.032459] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9c89af-6ab4-4393-bc99-94c28a68a32a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.042030] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aad8b26-7f5a-4fa1-9374-f36120c686e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.059802] env[63197]: DEBUG nova.compute.provider_tree [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.309357] env[63197]: DEBUG nova.network.neutron [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.479832] env[63197]: ERROR nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. [ 601.479832] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.479832] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.479832] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.479832] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.479832] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.479832] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.479832] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.479832] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.479832] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 601.479832] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.479832] env[63197]: ERROR nova.compute.manager raise self.value [ 601.479832] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.479832] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.479832] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.479832] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.480332] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.480332] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.480332] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. [ 601.480332] env[63197]: ERROR nova.compute.manager [ 601.480332] env[63197]: Traceback (most recent call last): [ 601.480332] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.480332] env[63197]: listener.cb(fileno) [ 601.480332] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.480332] env[63197]: result = function(*args, **kwargs) [ 601.480332] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.480332] env[63197]: return func(*args, **kwargs) [ 601.480332] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.480332] env[63197]: raise e [ 601.480332] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.480332] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 601.480332] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.480332] env[63197]: created_port_ids = self._update_ports_for_instance( [ 601.480332] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.480332] env[63197]: with excutils.save_and_reraise_exception(): [ 601.480332] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.480332] env[63197]: self.force_reraise() [ 601.480332] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.480332] env[63197]: raise self.value [ 601.480332] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.480332] env[63197]: updated_port = self._update_port( [ 601.480332] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.480332] env[63197]: _ensure_no_port_binding_failure(port) [ 601.480332] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.480332] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.481202] env[63197]: nova.exception.PortBindingFailed: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. [ 601.481202] env[63197]: Removing descriptor: 18 [ 601.481202] env[63197]: ERROR nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Traceback (most recent call last): [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] yield resources [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self.driver.spawn(context, instance, image_meta, [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.481202] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] vm_ref = self.build_virtual_machine(instance, [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] for vif in network_info: [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return self._sync_wrapper(fn, *args, **kwargs) [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self.wait() [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self[:] = self._gt.wait() [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return self._exit_event.wait() [ 601.481602] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] result = hub.switch() [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return self.greenlet.switch() [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] result = function(*args, **kwargs) [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return func(*args, **kwargs) [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] raise e [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] nwinfo = self.network_api.allocate_for_instance( [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.482055] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] created_port_ids = self._update_ports_for_instance( [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] with excutils.save_and_reraise_exception(): [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self.force_reraise() [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] raise self.value [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] updated_port = self._update_port( [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] _ensure_no_port_binding_failure(port) [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.482494] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] raise exception.PortBindingFailed(port_id=port['id']) [ 601.482857] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] nova.exception.PortBindingFailed: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. [ 601.482857] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] [ 601.482857] env[63197]: INFO nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Terminating instance [ 601.484116] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Acquiring lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.484291] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Acquired lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.484461] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.564442] env[63197]: DEBUG nova.scheduler.client.report [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.811333] env[63197]: INFO nova.compute.manager [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] Took 1.05 seconds to deallocate network for instance. [ 601.878122] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.878724] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.878724] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.878724] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.878983] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.878983] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.879985] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.880306] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.880563] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.880768] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.880977] env[63197]: DEBUG nova.virt.hardware [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.882195] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29320e7d-f8f1-452c-a4d4-0014ff1e5462 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.892190] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b606995-3a65-4b23-899c-5fc7ac54369d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.911760] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 601.917548] env[63197]: DEBUG oslo.service.loopingcall [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.917933] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 601.918198] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbc14f3c-2ca2-4602-b4e5-cf49007e6cea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.935430] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 601.935430] env[63197]: value = "task-1364043" [ 601.935430] env[63197]: _type = "Task" [ 601.935430] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.940735] env[63197]: DEBUG nova.compute.manager [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Received event network-changed-af208dde-7e08-4ea6-98b0-79296d2da0ae {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.940980] env[63197]: DEBUG nova.compute.manager [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Refreshing instance network info cache due to event network-changed-af208dde-7e08-4ea6-98b0-79296d2da0ae. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 601.941122] env[63197]: DEBUG oslo_concurrency.lockutils [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] Acquiring lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.946277] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364043, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.069955] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.070322] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.073437] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.923s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.194144] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.368672] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.451339] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364043, 'name': CreateVM_Task, 'duration_secs': 0.298275} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.451717] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 602.452305] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.452933] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.453560] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 602.454042] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26c6109c-40aa-4a0a-a59d-619420688d9a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.462019] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 602.462019] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e73cb0-8e08-c007-8f90-597cbc11bf44" [ 602.462019] env[63197]: _type = "Task" [ 602.462019] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 602.471504] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e73cb0-8e08-c007-8f90-597cbc11bf44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 602.578452] env[63197]: DEBUG nova.compute.utils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.579897] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.580164] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 602.690059] env[63197]: DEBUG nova.policy [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2f9aee5f7e2b472fb7852ed877bd933d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62e8a561d3f14bb9b236a2cc318b3ee4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 602.791370] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "920f5f9e-afb9-4ae6-a70b-5e902ea2c906" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.791626] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "920f5f9e-afb9-4ae6-a70b-5e902ea2c906" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.867358] env[63197]: INFO nova.scheduler.client.report [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Deleted allocations for instance 68ce48d7-e967-4cfc-a607-f6580bc14648 [ 602.872719] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Releasing lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.873149] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 602.873349] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 602.878651] env[63197]: DEBUG oslo_concurrency.lockutils [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] Acquired lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.878651] env[63197]: DEBUG nova.network.neutron [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Refreshing network info cache for port af208dde-7e08-4ea6-98b0-79296d2da0ae {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 602.878651] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfc6e252-8cb3-4b85-839c-bc88a586da11 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.888520] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b1cc6c-ca83-4ec7-b423-3a6beb9fe135 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.913351] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6581d2d-51e8-431f-97d6-acda6f438e9c could not be found. [ 602.914870] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 602.914870] env[63197]: INFO nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 602.914870] env[63197]: DEBUG oslo.service.loopingcall [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 602.914870] env[63197]: DEBUG nova.compute.manager [-] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.915147] env[63197]: DEBUG nova.network.neutron [-] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.954996] env[63197]: DEBUG nova.network.neutron [-] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.970672] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e73cb0-8e08-c007-8f90-597cbc11bf44, 'name': SearchDatastore_Task, 'duration_secs': 0.01072} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 602.971135] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.971575] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 602.972299] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.972683] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.973145] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 602.974171] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f94d21a9-7603-4146-92fe-cf40a4ca5e74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.986616] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 602.986616] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 602.986962] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55edc055-69a0-4ea0-a09d-4ed8707eafc7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.993747] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 602.993747] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529d3111-2f96-9b8f-9767-c3ea1b435d0d" [ 602.993747] env[63197]: _type = "Task" [ 602.993747] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.012390] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529d3111-2f96-9b8f-9767-c3ea1b435d0d, 'name': SearchDatastore_Task, 'duration_secs': 0.009016} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.013555] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-259d6d7f-7252-45bb-b21f-878f3c7b1dae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.019146] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 603.019146] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522e272e-2b37-3228-1311-4f3c4c0780ae" [ 603.019146] env[63197]: _type = "Task" [ 603.019146] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.034704] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522e272e-2b37-3228-1311-4f3c4c0780ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.066195] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5130155a-b03d-4016-b73b-9c2a749987a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.075934] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be870a4-1883-4703-86c2-c36d4743257a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.111897] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.116036] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e3a463-b49a-4c0d-92da-dd37eba0eda7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.124352] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9e5c2a-ad3b-4c88-b0f2-26e8c412ffd7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.139021] env[63197]: DEBUG nova.compute.provider_tree [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.382851] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d943b987-8bfd-4530-a098-f94769b3dfba tempest-TenantUsagesTestJSON-1391938219 tempest-TenantUsagesTestJSON-1391938219-project-member] Lock "68ce48d7-e967-4cfc-a607-f6580bc14648" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.898s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.386720] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "68ce48d7-e967-4cfc-a607-f6580bc14648" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.332s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.386928] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 68ce48d7-e967-4cfc-a607-f6580bc14648] During sync_power_state the instance has a pending task (spawning). Skip. [ 603.387113] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "68ce48d7-e967-4cfc-a607-f6580bc14648" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.428299] env[63197]: DEBUG nova.network.neutron [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.456721] env[63197]: DEBUG nova.network.neutron [-] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.531372] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522e272e-2b37-3228-1311-4f3c4c0780ae, 'name': SearchDatastore_Task, 'duration_secs': 0.009332} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.531987] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.532262] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 603.532520] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ec383f3-8b69-468a-9cd7-d8931b7d1cca {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.538756] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 603.538756] env[63197]: value = "task-1364044" [ 603.538756] env[63197]: _type = "Task" [ 603.538756] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.547108] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364044, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.555828] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Successfully created port: e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.624712] env[63197]: DEBUG nova.network.neutron [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.642481] env[63197]: DEBUG nova.scheduler.client.report [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.887338] env[63197]: DEBUG nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.961183] env[63197]: INFO nova.compute.manager [-] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Took 1.04 seconds to deallocate network for instance. [ 603.962689] env[63197]: DEBUG nova.compute.claims [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 603.963244] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.052248] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364044, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494472} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.052880] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 604.052880] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 604.053256] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca52340f-7e4f-4d00-b0cc-b79f8a90b6f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.062035] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 604.062035] env[63197]: value = "task-1364045" [ 604.062035] env[63197]: _type = "Task" [ 604.062035] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.072239] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364045, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.122739] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.128439] env[63197]: DEBUG oslo_concurrency.lockutils [req-0b46b35d-040e-4291-9487-edd2d6bcf0b6 req-d58d5e41-015c-4eee-986d-999d47c620d5 service nova] Releasing lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.149391] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.076s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.150157] env[63197]: ERROR nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Traceback (most recent call last): [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self.driver.spawn(context, instance, image_meta, [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] vm_ref = self.build_virtual_machine(instance, [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.150157] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] for vif in network_info: [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return self._sync_wrapper(fn, *args, **kwargs) [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self.wait() [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self[:] = self._gt.wait() [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return self._exit_event.wait() [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] result = hub.switch() [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.150778] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return self.greenlet.switch() [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] result = function(*args, **kwargs) [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] return func(*args, **kwargs) [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] raise e [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] nwinfo = self.network_api.allocate_for_instance( [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] created_port_ids = self._update_ports_for_instance( [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] with excutils.save_and_reraise_exception(): [ 604.151535] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] self.force_reraise() [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] raise self.value [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] updated_port = self._update_port( [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] _ensure_no_port_binding_failure(port) [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] raise exception.PortBindingFailed(port_id=port['id']) [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] nova.exception.PortBindingFailed: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. [ 604.153332] env[63197]: ERROR nova.compute.manager [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] [ 604.154171] env[63197]: DEBUG nova.compute.utils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.154171] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Build of instance 37280bd4-3cd8-4910-97de-0cc7494abc22 was re-scheduled: Binding failed for port e312e7b2-5270-4cf6-8678-7a1af450ad88, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.155347] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.155681] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Acquiring lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.156107] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Acquired lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.156107] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.161033] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.729s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.161033] env[63197]: INFO nova.compute.claims [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.169168] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.169624] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.169999] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.170763] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.170943] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.171114] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.171324] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.171495] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.171687] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.171860] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.172047] env[63197]: DEBUG nova.virt.hardware [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.172885] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800ec8a0-73a2-4e27-a2d7-556239255304 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.185369] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02196171-7493-48dc-820b-e57d11dc9854 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.411936] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.575414] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364045, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058298} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.575890] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 604.578154] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262a4aff-f923-4e5d-b34c-9acd877d6637 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.605425] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 604.606053] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66ff051f-9376-480f-a2ed-739c52d2b315 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.633503] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 604.633503] env[63197]: value = "task-1364046" [ 604.633503] env[63197]: _type = "Task" [ 604.633503] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 604.643894] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364046, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.689635] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.741671] env[63197]: DEBUG nova.compute.manager [req-10952c94-2b2d-495f-b73f-b74b349f8a6b req-84e6624c-5350-4903-af3e-0998d311e91f service nova] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Received event network-vif-deleted-af208dde-7e08-4ea6-98b0-79296d2da0ae {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.792656] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.148479] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364046, 'name': ReconfigVM_Task, 'duration_secs': 0.417007} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.148847] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 99539d7f-cd93-4250-97cd-2449e03e517d/99539d7f-cd93-4250-97cd-2449e03e517d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 605.149480] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-599bda75-d75c-4d2b-9f35-47af1c00063e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.158831] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 605.158831] env[63197]: value = "task-1364047" [ 605.158831] env[63197]: _type = "Task" [ 605.158831] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.169433] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364047, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.276211] env[63197]: ERROR nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. [ 605.276211] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.276211] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.276211] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.276211] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.276211] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.276211] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.276211] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.276211] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.276211] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 605.276211] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.276211] env[63197]: ERROR nova.compute.manager raise self.value [ 605.276211] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.276211] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.276211] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.276211] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.280144] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.280144] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.280144] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. [ 605.280144] env[63197]: ERROR nova.compute.manager [ 605.280144] env[63197]: Traceback (most recent call last): [ 605.280144] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.280144] env[63197]: listener.cb(fileno) [ 605.280144] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.280144] env[63197]: result = function(*args, **kwargs) [ 605.280144] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.280144] env[63197]: return func(*args, **kwargs) [ 605.280144] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.280144] env[63197]: raise e [ 605.280144] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.280144] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 605.280144] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.280144] env[63197]: created_port_ids = self._update_ports_for_instance( [ 605.280144] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.280144] env[63197]: with excutils.save_and_reraise_exception(): [ 605.280144] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.280144] env[63197]: self.force_reraise() [ 605.280144] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.280144] env[63197]: raise self.value [ 605.280144] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.280144] env[63197]: updated_port = self._update_port( [ 605.280144] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.280144] env[63197]: _ensure_no_port_binding_failure(port) [ 605.280144] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.280144] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.281222] env[63197]: nova.exception.PortBindingFailed: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. [ 605.281222] env[63197]: Removing descriptor: 18 [ 605.281222] env[63197]: ERROR nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Traceback (most recent call last): [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] yield resources [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self.driver.spawn(context, instance, image_meta, [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.281222] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] vm_ref = self.build_virtual_machine(instance, [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] for vif in network_info: [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return self._sync_wrapper(fn, *args, **kwargs) [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self.wait() [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self[:] = self._gt.wait() [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return self._exit_event.wait() [ 605.281640] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] result = hub.switch() [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return self.greenlet.switch() [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] result = function(*args, **kwargs) [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return func(*args, **kwargs) [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] raise e [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] nwinfo = self.network_api.allocate_for_instance( [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.282053] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] created_port_ids = self._update_ports_for_instance( [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] with excutils.save_and_reraise_exception(): [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self.force_reraise() [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] raise self.value [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] updated_port = self._update_port( [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] _ensure_no_port_binding_failure(port) [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.283106] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] raise exception.PortBindingFailed(port_id=port['id']) [ 605.283609] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] nova.exception.PortBindingFailed: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. [ 605.283609] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] [ 605.283609] env[63197]: INFO nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Terminating instance [ 605.283609] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Acquiring lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.283609] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Acquired lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.283926] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.303330] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Releasing lock "refresh_cache-37280bd4-3cd8-4910-97de-0cc7494abc22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.303566] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.303729] env[63197]: DEBUG nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.303892] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.322428] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.571482] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248d1bab-c340-4784-90b5-216a652d7e02 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.580495] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be017593-b4ad-4a53-8c07-e35faa9763d8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.618070] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10971558-f8d8-4e18-872a-2ad940588955 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.625447] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ea4c03-47e6-4efa-88f4-adac9b0514f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.644063] env[63197]: DEBUG nova.compute.provider_tree [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.671629] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364047, 'name': Rename_Task, 'duration_secs': 0.135722} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 605.671895] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 605.672138] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d44f56d-8390-4a93-adfb-f645353f8c74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.679298] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Waiting for the task: (returnval){ [ 605.679298] env[63197]: value = "task-1364048" [ 605.679298] env[63197]: _type = "Task" [ 605.679298] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 605.689101] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364048, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 605.810362] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.832372] env[63197]: DEBUG nova.network.neutron [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.923845] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.145332] env[63197]: DEBUG nova.scheduler.client.report [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.189514] env[63197]: DEBUG oslo_vmware.api [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Task: {'id': task-1364048, 'name': PowerOnVM_Task, 'duration_secs': 0.476639} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 606.189810] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 606.190590] env[63197]: DEBUG nova.compute.manager [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 606.191978] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e96cfe-0d84-4bd2-9bc6-89c4feb256bd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.338076] env[63197]: INFO nova.compute.manager [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] Took 1.03 seconds to deallocate network for instance. [ 606.426327] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Releasing lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.426861] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.427104] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.427461] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b35cb58-27e6-4eaa-917a-5cc4c3f958a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.438154] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b8ff34-a69f-4163-9c9e-6a8a05410a74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.454317] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "544ae361-619d-4e86-896c-21bf44e229d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.454548] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "544ae361-619d-4e86-896c-21bf44e229d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.466752] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2d8fb706-7400-43af-86fe-f90ac38393e0 could not be found. [ 606.466960] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.467164] env[63197]: INFO nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 606.467434] env[63197]: DEBUG oslo.service.loopingcall [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.469410] env[63197]: DEBUG nova.compute.manager [-] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.469410] env[63197]: DEBUG nova.network.neutron [-] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.490597] env[63197]: DEBUG nova.network.neutron [-] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.656456] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.656456] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.658331] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.204s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.713522] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.994695] env[63197]: DEBUG nova.network.neutron [-] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.166313] env[63197]: DEBUG nova.compute.utils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.174314] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.174314] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.255778] env[63197]: DEBUG nova.policy [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ea85355965034f05bf386bf88f8f04d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab847adc119a45c98db6c18fd27d0932', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.304526] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Acquiring lock "d7442338-8633-4866-89c4-1ef8e44004d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.304526] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Lock "d7442338-8633-4866-89c4-1ef8e44004d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.387960] env[63197]: DEBUG nova.compute.manager [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Received event network-changed-e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.388226] env[63197]: DEBUG nova.compute.manager [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Refreshing instance network info cache due to event network-changed-e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 607.388451] env[63197]: DEBUG oslo_concurrency.lockutils [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] Acquiring lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.388527] env[63197]: DEBUG oslo_concurrency.lockutils [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] Acquired lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.388671] env[63197]: DEBUG nova.network.neutron [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Refreshing network info cache for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 607.392911] env[63197]: INFO nova.scheduler.client.report [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Deleted allocations for instance 37280bd4-3cd8-4910-97de-0cc7494abc22 [ 607.498037] env[63197]: INFO nova.compute.manager [-] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Took 1.03 seconds to deallocate network for instance. [ 607.499636] env[63197]: DEBUG nova.compute.claims [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.499816] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.589703] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-016ab9e7-70d9-49b5-8c1b-abb4cc7d081e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.600151] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e3bd9e-d276-4c41-839d-b2a6a63750d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.643575] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1f20e2-17dd-4bf2-b7e5-348fddb29311 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.653251] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccb0aa4-09e5-4537-9aa5-29dcac2740cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.668907] env[63197]: DEBUG nova.compute.provider_tree [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.670960] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Successfully created port: a0c41c89-0eff-4df5-897a-3237572c810e {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.675303] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.906874] env[63197]: DEBUG oslo_concurrency.lockutils [None req-73fcc351-df43-41af-9f3e-d950b0f8a699 tempest-ServersAdminNegativeTestJSON-1114242303 tempest-ServersAdminNegativeTestJSON-1114242303-project-member] Lock "37280bd4-3cd8-4910-97de-0cc7494abc22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.101s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.908416] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "37280bd4-3cd8-4910-97de-0cc7494abc22" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 24.853s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.908600] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 37280bd4-3cd8-4910-97de-0cc7494abc22] During sync_power_state the instance has a pending task (spawning). Skip. [ 607.908770] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "37280bd4-3cd8-4910-97de-0cc7494abc22" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.929146] env[63197]: DEBUG nova.network.neutron [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.177455] env[63197]: DEBUG nova.scheduler.client.report [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.217958] env[63197]: DEBUG nova.network.neutron [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.273019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "99539d7f-cd93-4250-97cd-2449e03e517d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.273671] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "99539d7f-cd93-4250-97cd-2449e03e517d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.274039] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "99539d7f-cd93-4250-97cd-2449e03e517d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.274385] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "99539d7f-cd93-4250-97cd-2449e03e517d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.274750] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "99539d7f-cd93-4250-97cd-2449e03e517d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.277059] env[63197]: INFO nova.compute.manager [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Terminating instance [ 608.278850] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "refresh_cache-99539d7f-cd93-4250-97cd-2449e03e517d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.279088] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquired lock "refresh_cache-99539d7f-cd93-4250-97cd-2449e03e517d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.279286] env[63197]: DEBUG nova.network.neutron [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.410685] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.684334] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.026s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.685668] env[63197]: ERROR nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Traceback (most recent call last): [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self.driver.spawn(context, instance, image_meta, [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] vm_ref = self.build_virtual_machine(instance, [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.685668] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] for vif in network_info: [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return self._sync_wrapper(fn, *args, **kwargs) [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self.wait() [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self[:] = self._gt.wait() [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return self._exit_event.wait() [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] result = hub.switch() [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.686051] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return self.greenlet.switch() [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] result = function(*args, **kwargs) [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] return func(*args, **kwargs) [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] raise e [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] nwinfo = self.network_api.allocate_for_instance( [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] created_port_ids = self._update_ports_for_instance( [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] with excutils.save_and_reraise_exception(): [ 608.686426] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] self.force_reraise() [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] raise self.value [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] updated_port = self._update_port( [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] _ensure_no_port_binding_failure(port) [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] raise exception.PortBindingFailed(port_id=port['id']) [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] nova.exception.PortBindingFailed: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. [ 608.686857] env[63197]: ERROR nova.compute.manager [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] [ 608.688752] env[63197]: DEBUG nova.compute.utils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 608.688752] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.690951] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Build of instance a8315374-abb6-4663-bdb5-c4d88cdbbaa4 was re-scheduled: Binding failed for port 7b119857-88a0-4493-9c78-b94cd950e5f3, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 608.691129] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 608.691625] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.691707] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquired lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.693149] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.693149] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.757s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.726687] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:43:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1604493669',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-125340697',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.726954] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.727390] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.727390] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.728843] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.728843] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.728843] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.728843] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.728843] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.732030] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.732030] env[63197]: DEBUG nova.virt.hardware [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.732486] env[63197]: DEBUG oslo_concurrency.lockutils [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] Releasing lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.732569] env[63197]: DEBUG nova.compute.manager [req-7c0f63cf-8fcd-4272-9d26-696794535d77 req-49ba9c9a-a5c6-47fa-9a03-028c2bb81438 service nova] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Received event network-vif-deleted-e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.733968] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1964ca-63bc-40f1-bb35-be98596bf89f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.743681] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00bb7ccb-9a9a-4d15-8d53-6d9d91cf0e2d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.809140] env[63197]: DEBUG nova.network.neutron [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.895349] env[63197]: DEBUG nova.network.neutron [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.953296] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.222696] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.307104] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.397010] env[63197]: ERROR nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. [ 609.397010] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.397010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.397010] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.397010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.397010] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.397010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.397010] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.397010] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.397010] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 609.397010] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.397010] env[63197]: ERROR nova.compute.manager raise self.value [ 609.397010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.397010] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.397010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.397010] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.397701] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.397701] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.397701] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. [ 609.397701] env[63197]: ERROR nova.compute.manager [ 609.397701] env[63197]: Traceback (most recent call last): [ 609.397701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.397701] env[63197]: listener.cb(fileno) [ 609.397701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.397701] env[63197]: result = function(*args, **kwargs) [ 609.397701] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.397701] env[63197]: return func(*args, **kwargs) [ 609.397701] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.397701] env[63197]: raise e [ 609.397701] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.397701] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 609.397701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.397701] env[63197]: created_port_ids = self._update_ports_for_instance( [ 609.397701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.397701] env[63197]: with excutils.save_and_reraise_exception(): [ 609.397701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.397701] env[63197]: self.force_reraise() [ 609.397701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.397701] env[63197]: raise self.value [ 609.397701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.397701] env[63197]: updated_port = self._update_port( [ 609.397701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.397701] env[63197]: _ensure_no_port_binding_failure(port) [ 609.397701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.397701] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.398674] env[63197]: nova.exception.PortBindingFailed: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. [ 609.398674] env[63197]: Removing descriptor: 18 [ 609.398674] env[63197]: ERROR nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Traceback (most recent call last): [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] yield resources [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self.driver.spawn(context, instance, image_meta, [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.398674] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] vm_ref = self.build_virtual_machine(instance, [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] for vif in network_info: [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return self._sync_wrapper(fn, *args, **kwargs) [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self.wait() [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self[:] = self._gt.wait() [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return self._exit_event.wait() [ 609.399037] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] result = hub.switch() [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return self.greenlet.switch() [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] result = function(*args, **kwargs) [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return func(*args, **kwargs) [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] raise e [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] nwinfo = self.network_api.allocate_for_instance( [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.399479] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] created_port_ids = self._update_ports_for_instance( [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] with excutils.save_and_reraise_exception(): [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self.force_reraise() [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] raise self.value [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] updated_port = self._update_port( [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] _ensure_no_port_binding_failure(port) [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.399865] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] raise exception.PortBindingFailed(port_id=port['id']) [ 609.400366] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] nova.exception.PortBindingFailed: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. [ 609.400366] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] [ 609.400366] env[63197]: INFO nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Terminating instance [ 609.401081] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Releasing lock "refresh_cache-99539d7f-cd93-4250-97cd-2449e03e517d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.401164] env[63197]: DEBUG nova.compute.manager [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.401348] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 609.404581] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7b3a16-6cbf-4e62-96b1-5b69b3277bba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.408143] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Acquiring lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.409082] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Acquired lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.409082] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 609.414731] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 609.414997] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-189f49f9-6c87-44b2-8fc2-1d0812319e10 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.427017] env[63197]: DEBUG oslo_vmware.api [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 609.427017] env[63197]: value = "task-1364049" [ 609.427017] env[63197]: _type = "Task" [ 609.427017] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.438703] env[63197]: DEBUG oslo_vmware.api [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364049, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 609.593119] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1912e4ee-e321-40d9-8dd1-15633fc7ec33 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.601744] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333e3134-f263-4380-aeeb-26835101c7ce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.633807] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c2bd9f-c6dc-47f5-9e9b-b638957304c8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.641600] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6393ce2b-7a8e-45e0-8c0d-52c9d04269c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.656639] env[63197]: DEBUG nova.compute.provider_tree [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.811231] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Releasing lock "refresh_cache-a8315374-abb6-4663-bdb5-c4d88cdbbaa4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.811231] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 609.811511] env[63197]: DEBUG nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.811511] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.842529] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.936500] env[63197]: DEBUG nova.compute.manager [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Received event network-changed-a0c41c89-0eff-4df5-897a-3237572c810e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.936675] env[63197]: DEBUG nova.compute.manager [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Refreshing instance network info cache due to event network-changed-a0c41c89-0eff-4df5-897a-3237572c810e. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 609.936863] env[63197]: DEBUG oslo_concurrency.lockutils [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] Acquiring lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.937647] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.944637] env[63197]: DEBUG oslo_vmware.api [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364049, 'name': PowerOffVM_Task, 'duration_secs': 0.191208} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 609.945141] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 609.945323] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 609.945633] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49205490-c358-4cd0-81db-86b8b4705891 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.971452] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 609.971747] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 609.972379] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleting the datastore file [datastore2] 99539d7f-cd93-4250-97cd-2449e03e517d {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 609.972379] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5bf249c1-1c86-4a28-9db2-505e15a49f82 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.979049] env[63197]: DEBUG oslo_vmware.api [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for the task: (returnval){ [ 609.979049] env[63197]: value = "task-1364051" [ 609.979049] env[63197]: _type = "Task" [ 609.979049] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 609.988154] env[63197]: DEBUG oslo_vmware.api [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364051, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 610.102999] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.160649] env[63197]: DEBUG nova.scheduler.client.report [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.346754] env[63197]: DEBUG nova.network.neutron [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.494120] env[63197]: DEBUG oslo_vmware.api [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Task: {'id': task-1364051, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101244} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 610.494381] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 610.494554] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 610.495170] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.495379] env[63197]: INFO nova.compute.manager [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Took 1.09 seconds to destroy the instance on the hypervisor. [ 610.495616] env[63197]: DEBUG oslo.service.loopingcall [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.495805] env[63197]: DEBUG nova.compute.manager [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.495899] env[63197]: DEBUG nova.network.neutron [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.512322] env[63197]: DEBUG nova.network.neutron [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.608904] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Releasing lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.608904] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.608904] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 610.608904] env[63197]: DEBUG oslo_concurrency.lockutils [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] Acquired lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.608904] env[63197]: DEBUG nova.network.neutron [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Refreshing network info cache for port a0c41c89-0eff-4df5-897a-3237572c810e {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 610.609209] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21751000-9d73-4c06-95c2-b884fba53ab2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.622515] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c46d878-a1eb-409d-8247-8efe97403275 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.647463] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 441988d7-b5c1-4996-a6fa-4a6e21f86b22 could not be found. [ 610.647725] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.647907] env[63197]: INFO nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Took 0.04 seconds to destroy the instance on the hypervisor. [ 610.648376] env[63197]: DEBUG oslo.service.loopingcall [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.648376] env[63197]: DEBUG nova.compute.manager [-] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.648478] env[63197]: DEBUG nova.network.neutron [-] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 610.667076] env[63197]: DEBUG nova.network.neutron [-] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.670273] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.977s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.671243] env[63197]: ERROR nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Traceback (most recent call last): [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self.driver.spawn(context, instance, image_meta, [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] vm_ref = self.build_virtual_machine(instance, [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.671243] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] for vif in network_info: [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return self._sync_wrapper(fn, *args, **kwargs) [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self.wait() [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self[:] = self._gt.wait() [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return self._exit_event.wait() [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] result = hub.switch() [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.671796] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return self.greenlet.switch() [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] result = function(*args, **kwargs) [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] return func(*args, **kwargs) [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] raise e [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] nwinfo = self.network_api.allocate_for_instance( [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] created_port_ids = self._update_ports_for_instance( [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] with excutils.save_and_reraise_exception(): [ 610.672218] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] self.force_reraise() [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] raise self.value [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] updated_port = self._update_port( [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] _ensure_no_port_binding_failure(port) [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] raise exception.PortBindingFailed(port_id=port['id']) [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] nova.exception.PortBindingFailed: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. [ 610.672598] env[63197]: ERROR nova.compute.manager [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] [ 610.673144] env[63197]: DEBUG nova.compute.utils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.673144] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.540s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.674805] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Build of instance 06de0427-5e1a-469c-ae45-9dc72336d3cc was re-scheduled: Binding failed for port 8f7b4666-c370-4d86-85da-b99b1dcbc6db, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.675252] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.675478] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Acquiring lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.675625] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Acquired lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.675784] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.849620] env[63197]: INFO nova.compute.manager [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] Took 1.04 seconds to deallocate network for instance. [ 611.015928] env[63197]: DEBUG nova.network.neutron [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.135468] env[63197]: DEBUG nova.network.neutron [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.169751] env[63197]: DEBUG nova.network.neutron [-] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.219329] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.317159] env[63197]: DEBUG nova.network.neutron [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.384285] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.520596] env[63197]: INFO nova.compute.manager [-] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Took 1.02 seconds to deallocate network for instance. [ 611.672601] env[63197]: INFO nova.compute.manager [-] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Took 1.02 seconds to deallocate network for instance. [ 611.677807] env[63197]: DEBUG nova.compute.claims [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.677807] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.717675] env[63197]: WARNING nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 611.717844] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 99539d7f-cd93-4250-97cd-2449e03e517d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 611.820875] env[63197]: DEBUG oslo_concurrency.lockutils [req-ffeabb03-a4a8-4251-85e5-7cc20692676e req-75d99e0b-ae2a-4e8b-970a-5bf4f7904209 service nova] Releasing lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.887640] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Releasing lock "refresh_cache-06de0427-5e1a-469c-ae45-9dc72336d3cc" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.887866] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.888057] env[63197]: DEBUG nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.888221] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.895142] env[63197]: INFO nova.scheduler.client.report [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Deleted allocations for instance a8315374-abb6-4663-bdb5-c4d88cdbbaa4 [ 611.916127] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.028620] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.221733] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a8315374-abb6-4663-bdb5-c4d88cdbbaa4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 612.293744] env[63197]: DEBUG nova.compute.manager [req-e98fecb3-9c2a-4408-88b4-9084ac6a84b0 req-c366dfee-e4c2-4fe1-b280-4af9c4f34a51 service nova] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Received event network-vif-deleted-a0c41c89-0eff-4df5-897a-3237572c810e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.410897] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e535af6-e1e1-4957-acf0-b9e0bc412414 tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "a8315374-abb6-4663-bdb5-c4d88cdbbaa4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.231s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.412253] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "a8315374-abb6-4663-bdb5-c4d88cdbbaa4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 29.357s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.415022] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-320d81e5-4b10-425c-bbcc-9f701deb9650 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.421696] env[63197]: DEBUG nova.network.neutron [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.429787] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b60654-29ca-4d43-9117-09a13d4b5199 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.726739] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 06de0427-5e1a-469c-ae45-9dc72336d3cc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 612.726919] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance b6581d2d-51e8-431f-97d6-acda6f438e9c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 612.727068] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 2d8fb706-7400-43af-86fe-f90ac38393e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 612.727193] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 441988d7-b5c1-4996-a6fa-4a6e21f86b22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 612.913872] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.947092] env[63197]: INFO nova.compute.manager [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] [instance: 06de0427-5e1a-469c-ae45-9dc72336d3cc] Took 1.06 seconds to deallocate network for instance. [ 612.961272] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a8315374-abb6-4663-bdb5-c4d88cdbbaa4] During the sync_power process the instance has moved from host None to host cpu-1 [ 612.962280] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "a8315374-abb6-4663-bdb5-c4d88cdbbaa4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.549s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.231622] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance bb7f3278-8678-4053-b957-9ceee430089d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 613.444907] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.736225] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 614.013198] env[63197]: INFO nova.scheduler.client.report [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Deleted allocations for instance 06de0427-5e1a-469c-ae45-9dc72336d3cc [ 614.240256] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance d84f11a3-ed86-4fc4-9092-0babc7fccb3f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 614.524680] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7226f194-269e-4a37-aa31-30ac7de0de62 tempest-FloatingIPsAssociationNegativeTestJSON-84255563 tempest-FloatingIPsAssociationNegativeTestJSON-84255563-project-member] Lock "06de0427-5e1a-469c-ae45-9dc72336d3cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.238s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.751434] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a066e6e1-12a4-455b-87cc-e3d6a6a17c1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 615.029081] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.253155] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance f01a025d-2cb9-4cb9-878a-b05e318c21b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 615.554025] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.759111] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance b52f3d4b-7699-484d-8b4e-84393848fb37 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 616.262052] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a18802fb-2a73-4bb5-aab9-5a7eb599a5a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 616.766898] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 47da4c8f-a4cb-4d7f-81c0-7c3019169b56 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 617.274327] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 4f2d4315-d533-431a-970a-ea655febf0ab has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 617.764392] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.764625] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.781955] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 6abf52eb-a394-477f-95e6-87cddecd4bd0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 618.287012] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 0d7e1031-896b-4ec3-9da0-74d65b1ef78a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 618.791030] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 11c6fcd4-87d3-4ceb-98eb-a46ca2910584 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 619.296831] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 68b27a14-f52b-4d89-9cdc-be0529428eb7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 619.802152] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 920f5f9e-afb9-4ae6-a70b-5e902ea2c906 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 620.306860] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 544ae361-619d-4e86-896c-21bf44e229d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 620.813562] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance d7442338-8633-4866-89c4-1ef8e44004d2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 620.813562] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 620.813562] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 621.241427] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c403314-694c-4c2e-9303-10876fe58d56 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.249166] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d0b679-e18e-4b2e-a464-c6f0ee0fa7ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.288106] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96493bcf-0148-4760-8269-35d0f9673f38 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.297239] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d529906-c3ae-4878-88d3-9c9a22a40f4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.311553] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.817754] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.053794] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Acquiring lock "1e650df0-b4a3-45dd-8b14-b41d4c77b625" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.054072] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Lock "1e650df0-b4a3-45dd-8b14-b41d4c77b625" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.322370] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 622.323076] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.650s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.323076] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.795s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.323076] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.325638] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.324s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.327234] env[63197]: INFO nova.compute.claims [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.329661] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.330038] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Getting list of instances from cluster (obj){ [ 622.330038] env[63197]: value = "domain-c8" [ 622.330038] env[63197]: _type = "ClusterComputeResource" [ 622.330038] env[63197]: } {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 622.331430] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e3ad8e-44b5-4c73-9747-0fd314db99dd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.348369] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Got total of 0 instances {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 622.363386] env[63197]: INFO nova.scheduler.client.report [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Deleted allocations for instance 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6 [ 622.872369] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ac52ba77-cc04-4dce-85c6-6f83959c8d85 tempest-ServerDiagnosticsV248Test-615614865 tempest-ServerDiagnosticsV248Test-615614865-project-member] Lock "9c27469a-b8f9-4374-ae2a-ca17ed43f4f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.164s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.741110] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d981d2-33c3-4fd6-a6c5-327f26af4796 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.749273] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fc2d6e-1ae4-44ed-8fd9-a0de0a4aab95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.796184] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e583cd21-fa43-40ee-ac2f-d8542bf3b048 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.805807] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669c4f38-1a88-4119-ad5c-92961f48be92 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.825335] env[63197]: DEBUG nova.compute.provider_tree [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.330198] env[63197]: DEBUG nova.scheduler.client.report [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.810067] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "09dc9549-ab9a-4f36-9671-ab1b5595649d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.810355] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "09dc9549-ab9a-4f36-9671-ab1b5595649d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.838467] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.839022] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.841959] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.322s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.841959] env[63197]: DEBUG nova.objects.instance [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63197) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 625.348205] env[63197]: DEBUG nova.compute.utils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.352412] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.352514] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 625.529039] env[63197]: DEBUG nova.policy [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fdfd1462469a4bdb9fe2e0240701eae7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a356813929b4d0eb1129de474ddaffe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.853427] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.861575] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2e634108-c9db-4a14-abd6-6ef1b954c91e tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.863119] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.939s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.864842] env[63197]: INFO nova.compute.claims [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.282968] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Successfully created port: 578d90c4-64f2-4194-9d29-f171e09a6c41 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.872307] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.920369] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.920665] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.920771] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.920953] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.921107] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.926241] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.926241] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.926241] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.926241] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.926241] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.926444] env[63197]: DEBUG nova.virt.hardware [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.926444] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa04e2a-b031-4a38-8c97-42e0a317ecb6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.936537] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "0f566f8c-c92a-4885-94a6-08043857a1c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.936693] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "0f566f8c-c92a-4885-94a6-08043857a1c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.941901] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63076347-7e77-46a6-a069-96b1c30a861e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.344071] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b78dc1-4fa0-46c9-bfa1-8e3074ddce9a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.352915] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a27f651-b9b0-4dca-9307-d233abe0a4a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.388415] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a83519-a6a7-476a-b75b-f5095fac6273 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.396393] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c62b88-83bb-4c69-9c77-7416b8e8d82c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.422656] env[63197]: DEBUG nova.compute.provider_tree [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.722278] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "a0638d5a-6d33-4eb6-a9b6-ad6620219259" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.722539] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "a0638d5a-6d33-4eb6-a9b6-ad6620219259" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.926667] env[63197]: DEBUG nova.scheduler.client.report [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.436356] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.437107] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 628.439993] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.456s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.441595] env[63197]: INFO nova.compute.claims [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.594978] env[63197]: DEBUG nova.compute.manager [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Received event network-changed-578d90c4-64f2-4194-9d29-f171e09a6c41 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.594978] env[63197]: DEBUG nova.compute.manager [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Refreshing instance network info cache due to event network-changed-578d90c4-64f2-4194-9d29-f171e09a6c41. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 628.594978] env[63197]: DEBUG oslo_concurrency.lockutils [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] Acquiring lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.594978] env[63197]: DEBUG oslo_concurrency.lockutils [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] Acquired lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.594978] env[63197]: DEBUG nova.network.neutron [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Refreshing network info cache for port 578d90c4-64f2-4194-9d29-f171e09a6c41 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 628.792990] env[63197]: ERROR nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. [ 628.792990] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.792990] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.792990] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.792990] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.792990] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.792990] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.792990] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.792990] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.792990] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 628.792990] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.792990] env[63197]: ERROR nova.compute.manager raise self.value [ 628.792990] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.792990] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.792990] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.792990] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.793791] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.793791] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.793791] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. [ 628.793791] env[63197]: ERROR nova.compute.manager [ 628.793791] env[63197]: Traceback (most recent call last): [ 628.793791] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.793791] env[63197]: listener.cb(fileno) [ 628.793791] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.793791] env[63197]: result = function(*args, **kwargs) [ 628.793791] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.793791] env[63197]: return func(*args, **kwargs) [ 628.793791] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.793791] env[63197]: raise e [ 628.793791] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.793791] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 628.793791] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.793791] env[63197]: created_port_ids = self._update_ports_for_instance( [ 628.793791] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.793791] env[63197]: with excutils.save_and_reraise_exception(): [ 628.793791] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.793791] env[63197]: self.force_reraise() [ 628.793791] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.793791] env[63197]: raise self.value [ 628.793791] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.793791] env[63197]: updated_port = self._update_port( [ 628.793791] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.793791] env[63197]: _ensure_no_port_binding_failure(port) [ 628.793791] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.793791] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.796350] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. [ 628.796350] env[63197]: Removing descriptor: 17 [ 628.796350] env[63197]: ERROR nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] Traceback (most recent call last): [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] yield resources [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self.driver.spawn(context, instance, image_meta, [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.796350] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] vm_ref = self.build_virtual_machine(instance, [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] for vif in network_info: [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return self._sync_wrapper(fn, *args, **kwargs) [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self.wait() [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self[:] = self._gt.wait() [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return self._exit_event.wait() [ 628.796990] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] result = hub.switch() [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return self.greenlet.switch() [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] result = function(*args, **kwargs) [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return func(*args, **kwargs) [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] raise e [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] nwinfo = self.network_api.allocate_for_instance( [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.797534] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] created_port_ids = self._update_ports_for_instance( [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] with excutils.save_and_reraise_exception(): [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self.force_reraise() [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] raise self.value [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] updated_port = self._update_port( [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] _ensure_no_port_binding_failure(port) [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.797921] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] raise exception.PortBindingFailed(port_id=port['id']) [ 628.798407] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] nova.exception.PortBindingFailed: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. [ 628.798407] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] [ 628.798407] env[63197]: INFO nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Terminating instance [ 628.798407] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.946101] env[63197]: DEBUG nova.compute.utils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.947084] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.947282] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 628.989585] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Acquiring lock "715c53a9-224e-4457-9df9-a2ada9fcae93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.989741] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Lock "715c53a9-224e-4457-9df9-a2ada9fcae93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.047441] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "86a728c9-27d3-4c98-a4d6-6b0736f24648" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.047441] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "86a728c9-27d3-4c98-a4d6-6b0736f24648" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.071021] env[63197]: DEBUG nova.policy [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45050bfa90e944978eb460d9d52f5e9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a5d1b44abfa4b449310d025aa8a0ac7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 629.135944] env[63197]: DEBUG nova.network.neutron [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.324132] env[63197]: DEBUG nova.network.neutron [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.457726] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.702218] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Successfully created port: e8b871f2-4263-48c9-9202-655d88ff6865 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.826874] env[63197]: DEBUG oslo_concurrency.lockutils [req-bac7e7d9-6f76-4676-b6a1-b5b3ec6b773e req-4ffdb4e0-eb3e-4b8d-bb67-252ae8c9052b service nova] Releasing lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.828527] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquired lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.828527] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.923136] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe97aef9-4bf0-4e77-9749-56aa8fc22f33 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.933305] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e96eab-bb1d-4916-8243-26d7bf1ed7bc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.969726] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a365235b-109c-4fb6-9e79-df1f77eddc45 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.982071] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02907287-4d1c-4423-b4ef-a23aac658e37 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.995358] env[63197]: DEBUG nova.compute.provider_tree [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.464394] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.475383] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 630.503493] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 630.503872] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 630.504068] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 630.504260] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 630.504403] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 630.505135] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 630.505135] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 630.505135] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 630.506619] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 630.506856] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 630.507060] env[63197]: DEBUG nova.virt.hardware [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.508095] env[63197]: DEBUG nova.scheduler.client.report [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.512518] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44deffba-851f-40f8-9000-164a7610431e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.526566] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926e6c5d-4a32-48d0-ab5e-ef36e9fd6a06 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.650400] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.018869] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.019223] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 631.025552] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.061s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.035906] env[63197]: DEBUG nova.compute.manager [req-d2129c06-ef8b-4214-b520-4c6543813031 req-00850c0e-e1e1-4088-ba4e-1cbbb0f1eb82 service nova] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Received event network-vif-deleted-578d90c4-64f2-4194-9d29-f171e09a6c41 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.156239] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Releasing lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.156239] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.156239] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.156239] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73d6f1b6-4503-4359-a87c-902b31e39541 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.164791] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d41b3d7-e8ce-4b07-b383-a2e85838b622 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.188827] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bb7f3278-8678-4053-b957-9ceee430089d could not be found. [ 631.189135] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.189367] env[63197]: INFO nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 631.189676] env[63197]: DEBUG oslo.service.loopingcall [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.189942] env[63197]: DEBUG nova.compute.manager [-] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.190144] env[63197]: DEBUG nova.network.neutron [-] [instance: bb7f3278-8678-4053-b957-9ceee430089d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.210273] env[63197]: DEBUG nova.network.neutron [-] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.477258] env[63197]: ERROR nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. [ 631.477258] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.477258] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.477258] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.477258] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.477258] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.477258] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.477258] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.477258] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.477258] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 631.477258] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.477258] env[63197]: ERROR nova.compute.manager raise self.value [ 631.477258] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.477258] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.477258] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.477258] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.477922] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.477922] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.477922] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. [ 631.477922] env[63197]: ERROR nova.compute.manager [ 631.477922] env[63197]: Traceback (most recent call last): [ 631.477922] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.477922] env[63197]: listener.cb(fileno) [ 631.477922] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.477922] env[63197]: result = function(*args, **kwargs) [ 631.477922] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.477922] env[63197]: return func(*args, **kwargs) [ 631.477922] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.477922] env[63197]: raise e [ 631.477922] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.477922] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 631.477922] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.477922] env[63197]: created_port_ids = self._update_ports_for_instance( [ 631.477922] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.477922] env[63197]: with excutils.save_and_reraise_exception(): [ 631.477922] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.477922] env[63197]: self.force_reraise() [ 631.477922] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.477922] env[63197]: raise self.value [ 631.477922] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.477922] env[63197]: updated_port = self._update_port( [ 631.477922] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.477922] env[63197]: _ensure_no_port_binding_failure(port) [ 631.477922] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.477922] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.478859] env[63197]: nova.exception.PortBindingFailed: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. [ 631.478859] env[63197]: Removing descriptor: 17 [ 631.478859] env[63197]: ERROR nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Traceback (most recent call last): [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] yield resources [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self.driver.spawn(context, instance, image_meta, [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.478859] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] vm_ref = self.build_virtual_machine(instance, [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] for vif in network_info: [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return self._sync_wrapper(fn, *args, **kwargs) [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self.wait() [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self[:] = self._gt.wait() [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return self._exit_event.wait() [ 631.479362] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] result = hub.switch() [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return self.greenlet.switch() [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] result = function(*args, **kwargs) [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return func(*args, **kwargs) [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] raise e [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] nwinfo = self.network_api.allocate_for_instance( [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.480119] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] created_port_ids = self._update_ports_for_instance( [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] with excutils.save_and_reraise_exception(): [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self.force_reraise() [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] raise self.value [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] updated_port = self._update_port( [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] _ensure_no_port_binding_failure(port) [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.480710] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] raise exception.PortBindingFailed(port_id=port['id']) [ 631.482485] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] nova.exception.PortBindingFailed: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. [ 631.482485] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] [ 631.482485] env[63197]: INFO nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Terminating instance [ 631.482485] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Acquiring lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.482485] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Acquired lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.482485] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.522313] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Acquiring lock "218ad539-e2d0-4d4b-9857-ec9d2af0be03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.522313] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Lock "218ad539-e2d0-4d4b-9857-ec9d2af0be03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.533375] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Acquiring lock "d9f1dcc1-77fd-4c8c-b930-eb421691dc94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.533658] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Lock "d9f1dcc1-77fd-4c8c-b930-eb421691dc94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.536374] env[63197]: DEBUG nova.compute.utils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.539702] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.539924] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 631.608850] env[63197]: DEBUG nova.policy [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fdfd1462469a4bdb9fe2e0240701eae7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a356813929b4d0eb1129de474ddaffe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.712092] env[63197]: DEBUG nova.network.neutron [-] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.932852] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Successfully created port: 8bee1fbe-7a99-4330-af05-7834adb33c6e {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.023726] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.031156] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79563b2e-1d7f-4822-8385-3786466c7f55 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.040138] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.044990] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f18a149-c2c6-4430-96e1-a68523623225 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.076581] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f9dc8d-3a23-4e62-a6c2-4c25f4a5d04a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.085287] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ca9042-1184-4c8d-9443-4292c973710c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.099069] env[63197]: DEBUG nova.compute.provider_tree [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.174547] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.218478] env[63197]: INFO nova.compute.manager [-] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Took 1.03 seconds to deallocate network for instance. [ 632.223890] env[63197]: DEBUG nova.compute.claims [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.224093] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.604447] env[63197]: DEBUG nova.scheduler.client.report [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.677482] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Releasing lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.678037] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.678132] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.678418] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70df21c3-85ac-4b3d-aedc-4d34d1a2b7b0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.695155] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baadc585-f927-40ef-a001-e4c0d2cc2e22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.719563] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282 could not be found. [ 632.719563] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.719563] env[63197]: INFO nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Took 0.04 seconds to destroy the instance on the hypervisor. [ 632.719563] env[63197]: DEBUG oslo.service.loopingcall [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.719563] env[63197]: DEBUG nova.compute.manager [-] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.719563] env[63197]: DEBUG nova.network.neutron [-] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.741760] env[63197]: DEBUG nova.network.neutron [-] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.052973] env[63197]: ERROR nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. [ 633.052973] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.052973] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.052973] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.052973] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.052973] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.052973] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.052973] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.052973] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.052973] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 633.052973] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.052973] env[63197]: ERROR nova.compute.manager raise self.value [ 633.052973] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.052973] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.052973] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.052973] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.053836] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.053836] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.053836] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. [ 633.053836] env[63197]: ERROR nova.compute.manager [ 633.053836] env[63197]: Traceback (most recent call last): [ 633.053836] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.053836] env[63197]: listener.cb(fileno) [ 633.053836] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.053836] env[63197]: result = function(*args, **kwargs) [ 633.053836] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.053836] env[63197]: return func(*args, **kwargs) [ 633.053836] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.053836] env[63197]: raise e [ 633.053836] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.053836] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 633.053836] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.053836] env[63197]: created_port_ids = self._update_ports_for_instance( [ 633.053836] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.053836] env[63197]: with excutils.save_and_reraise_exception(): [ 633.053836] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.053836] env[63197]: self.force_reraise() [ 633.053836] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.053836] env[63197]: raise self.value [ 633.053836] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.053836] env[63197]: updated_port = self._update_port( [ 633.053836] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.053836] env[63197]: _ensure_no_port_binding_failure(port) [ 633.053836] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.053836] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.054841] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. [ 633.054841] env[63197]: Removing descriptor: 17 [ 633.056265] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.086323] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.086692] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.086948] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.087446] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.087822] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.087902] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.088198] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.089649] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.089649] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.089649] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.089649] env[63197]: DEBUG nova.virt.hardware [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.090359] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4489bd0-a3ac-4904-b4a3-3cce13f3e201 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.098146] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c73e882-922f-4b46-926d-4a64ae222a65 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.104497] env[63197]: DEBUG nova.compute.manager [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Received event network-changed-e8b871f2-4263-48c9-9202-655d88ff6865 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.104677] env[63197]: DEBUG nova.compute.manager [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Refreshing instance network info cache due to event network-changed-e8b871f2-4263-48c9-9202-655d88ff6865. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 633.104887] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] Acquiring lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.105051] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] Acquired lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.105266] env[63197]: DEBUG nova.network.neutron [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Refreshing network info cache for port e8b871f2-4263-48c9-9202-655d88ff6865 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 633.119019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.093s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.119019] env[63197]: ERROR nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. [ 633.119019] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Traceback (most recent call last): [ 633.119019] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.119019] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self.driver.spawn(context, instance, image_meta, [ 633.119019] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 633.119019] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.119019] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.119019] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] vm_ref = self.build_virtual_machine(instance, [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] for vif in network_info: [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return self._sync_wrapper(fn, *args, **kwargs) [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self.wait() [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self[:] = self._gt.wait() [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return self._exit_event.wait() [ 633.119447] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] result = hub.switch() [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return self.greenlet.switch() [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] result = function(*args, **kwargs) [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] return func(*args, **kwargs) [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] raise e [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] nwinfo = self.network_api.allocate_for_instance( [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.119966] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] created_port_ids = self._update_ports_for_instance( [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] with excutils.save_and_reraise_exception(): [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] self.force_reraise() [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] raise self.value [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] updated_port = self._update_port( [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] _ensure_no_port_binding_failure(port) [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.120436] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] raise exception.PortBindingFailed(port_id=port['id']) [ 633.120875] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] nova.exception.PortBindingFailed: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. [ 633.120875] env[63197]: ERROR nova.compute.manager [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] [ 633.124019] env[63197]: DEBUG nova.compute.utils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 633.124019] env[63197]: ERROR nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. [ 633.124019] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Traceback (most recent call last): [ 633.124019] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 633.124019] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] yield resources [ 633.124019] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.124019] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self.driver.spawn(context, instance, image_meta, [ 633.124019] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 633.124019] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] vm_ref = self.build_virtual_machine(instance, [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] for vif in network_info: [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] return self._sync_wrapper(fn, *args, **kwargs) [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self.wait() [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self[:] = self._gt.wait() [ 633.124420] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] return self._exit_event.wait() [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] current.throw(*self._exc) [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] result = function(*args, **kwargs) [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] return func(*args, **kwargs) [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] raise e [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] nwinfo = self.network_api.allocate_for_instance( [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.124886] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] created_port_ids = self._update_ports_for_instance( [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] with excutils.save_and_reraise_exception(): [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self.force_reraise() [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] raise self.value [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] updated_port = self._update_port( [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] _ensure_no_port_binding_failure(port) [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.125346] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] raise exception.PortBindingFailed(port_id=port['id']) [ 633.125838] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] nova.exception.PortBindingFailed: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. [ 633.125838] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] [ 633.125838] env[63197]: INFO nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Terminating instance [ 633.127298] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Build of instance b6581d2d-51e8-431f-97d6-acda6f438e9c was re-scheduled: Binding failed for port af208dde-7e08-4ea6-98b0-79296d2da0ae, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 633.127754] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 633.127985] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Acquiring lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.128146] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Acquired lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.128304] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.129265] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.717s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.130780] env[63197]: INFO nova.compute.claims [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.139635] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.139635] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquired lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.139635] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.245476] env[63197]: DEBUG nova.network.neutron [-] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.625158] env[63197]: DEBUG nova.network.neutron [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.660122] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.662867] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.748152] env[63197]: INFO nova.compute.manager [-] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Took 1.03 seconds to deallocate network for instance. [ 633.752115] env[63197]: DEBUG nova.compute.claims [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.752181] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.754687] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.769397] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.771045] env[63197]: DEBUG nova.network.neutron [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.256549] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Releasing lock "refresh_cache-b6581d2d-51e8-431f-97d6-acda6f438e9c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.256872] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 634.256936] env[63197]: DEBUG nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.257099] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.275969] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Releasing lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.276431] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.276554] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.277091] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] Releasing lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.277301] env[63197]: DEBUG nova.compute.manager [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Received event network-vif-deleted-e8b871f2-4263-48c9-9202-655d88ff6865 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.277474] env[63197]: DEBUG nova.compute.manager [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Received event network-changed-8bee1fbe-7a99-4330-af05-7834adb33c6e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.277638] env[63197]: DEBUG nova.compute.manager [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Refreshing instance network info cache due to event network-changed-8bee1fbe-7a99-4330-af05-7834adb33c6e. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 634.277849] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] Acquiring lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.277968] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] Acquired lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.278142] env[63197]: DEBUG nova.network.neutron [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Refreshing network info cache for port 8bee1fbe-7a99-4330-af05-7834adb33c6e {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 634.279640] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.280654] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f67c742-f86f-44a2-8e80-029b30c6fcfa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.299573] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cafee62-eec9-46d8-ac1a-ead5777c40b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.325757] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d84f11a3-ed86-4fc4-9092-0babc7fccb3f could not be found. [ 634.326150] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.326644] env[63197]: INFO nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 634.327016] env[63197]: DEBUG oslo.service.loopingcall [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.329596] env[63197]: DEBUG nova.compute.manager [-] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.329841] env[63197]: DEBUG nova.network.neutron [-] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.351257] env[63197]: DEBUG nova.network.neutron [-] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.556239] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb4189d-da77-4515-9293-0314b1895511 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.564544] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a2bcfd-6e16-431f-9884-4908e8c11e24 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.598435] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5c88f8-3c18-49da-8dc7-5d916df14dcd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.606028] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdaf074-0cdc-4cfb-b51f-aef9c1835fe2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.623156] env[63197]: DEBUG nova.compute.provider_tree [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.786693] env[63197]: DEBUG nova.network.neutron [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.803751] env[63197]: DEBUG nova.network.neutron [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.856059] env[63197]: DEBUG nova.network.neutron [-] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.894390] env[63197]: DEBUG nova.network.neutron [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.126687] env[63197]: DEBUG nova.scheduler.client.report [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.131230] env[63197]: DEBUG nova.compute.manager [req-c1f1f9c7-e8c0-4e1b-ba28-a422ac05725a req-cd266b4f-78ef-4daa-952c-51a0742f75c9 service nova] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Received event network-vif-deleted-8bee1fbe-7a99-4330-af05-7834adb33c6e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.291137] env[63197]: INFO nova.compute.manager [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] [instance: b6581d2d-51e8-431f-97d6-acda6f438e9c] Took 1.03 seconds to deallocate network for instance. [ 635.359535] env[63197]: INFO nova.compute.manager [-] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Took 1.03 seconds to deallocate network for instance. [ 635.362210] env[63197]: DEBUG nova.compute.claims [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.362401] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.398431] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ee08f1c-beed-47a9-b43b-6c280f8602d5 req-9fc2615b-4339-406c-88b8-8cc55e909fec service nova] Releasing lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.633218] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.633746] env[63197]: DEBUG nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.636278] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.923s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.637163] env[63197]: DEBUG nova.objects.instance [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63197) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 636.141298] env[63197]: DEBUG nova.compute.utils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.145171] env[63197]: DEBUG nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Not allocating networking since 'none' was specified. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 636.323920] env[63197]: INFO nova.scheduler.client.report [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Deleted allocations for instance b6581d2d-51e8-431f-97d6-acda6f438e9c [ 636.646977] env[63197]: DEBUG nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 636.650286] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b3c7b15c-8c58-4ea0-9481-25b4c76609cb tempest-ServersAdmin275Test-1909151257 tempest-ServersAdmin275Test-1909151257-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.651275] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.151s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.831773] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4f7c4a64-926b-4499-ad5e-cbf895910859 tempest-ServerDiagnosticsNegativeTest-1867223677 tempest-ServerDiagnosticsNegativeTest-1867223677-project-member] Lock "b6581d2d-51e8-431f-97d6-acda6f438e9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.461s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.338569] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 637.604655] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58506cb-ad2f-4986-b940-133cd30580bf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.615578] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e08f173-ea6c-4985-a0a6-4cef3e1eda4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.657879] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6864ee-232c-499a-a091-5ec35cf70010 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.664512] env[63197]: DEBUG nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.672703] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4876659-ba16-4d5f-a256-513784aa84f0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.689163] env[63197]: DEBUG nova.compute.provider_tree [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.707725] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.707980] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.708180] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.708368] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.708535] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.708694] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.708922] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.709116] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.709495] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.709495] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.709680] env[63197]: DEBUG nova.virt.hardware [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.710862] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b945095a-dc78-4532-9efc-c8373723022e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.718949] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6a1e14-546a-400e-bd55-6eeb5df53530 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.732691] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 637.738658] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Creating folder: Project (b5454fb837c94f8ca9557cda4725ce04). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.738935] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-744875d3-6cc9-4947-b3ea-c5514d9935c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.750436] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Created folder: Project (b5454fb837c94f8ca9557cda4725ce04) in parent group-v290286. [ 637.750621] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Creating folder: Instances. Parent ref: group-v290299. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.751361] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2194335-1711-4c22-930e-64d3a61c9b75 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.760227] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Created folder: Instances in parent group-v290299. [ 637.760454] env[63197]: DEBUG oslo.service.loopingcall [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 637.760637] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 637.760847] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5598840e-6ac5-4b5b-b2d7-ceddbab424fb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.777884] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 637.777884] env[63197]: value = "task-1364062" [ 637.777884] env[63197]: _type = "Task" [ 637.777884] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.784683] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364062, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.858473] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.193018] env[63197]: DEBUG nova.scheduler.client.report [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.288352] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364062, 'name': CreateVM_Task, 'duration_secs': 0.26765} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.288557] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 638.289354] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.289354] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.289775] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 638.290541] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6258489b-7c59-4785-a98a-a2f946970374 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.294311] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 638.294311] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b79794-454a-0f15-6b5e-5c5783510be1" [ 638.294311] env[63197]: _type = "Task" [ 638.294311] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.301600] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b79794-454a-0f15-6b5e-5c5783510be1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.697652] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.046s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.698321] env[63197]: ERROR nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Traceback (most recent call last): [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self.driver.spawn(context, instance, image_meta, [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] vm_ref = self.build_virtual_machine(instance, [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.698321] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] for vif in network_info: [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return self._sync_wrapper(fn, *args, **kwargs) [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self.wait() [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self[:] = self._gt.wait() [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return self._exit_event.wait() [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] result = hub.switch() [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.698558] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return self.greenlet.switch() [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] result = function(*args, **kwargs) [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] return func(*args, **kwargs) [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] raise e [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] nwinfo = self.network_api.allocate_for_instance( [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] created_port_ids = self._update_ports_for_instance( [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] with excutils.save_and_reraise_exception(): [ 638.698829] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] self.force_reraise() [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] raise self.value [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] updated_port = self._update_port( [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] _ensure_no_port_binding_failure(port) [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] raise exception.PortBindingFailed(port_id=port['id']) [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] nova.exception.PortBindingFailed: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. [ 638.699092] env[63197]: ERROR nova.compute.manager [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] [ 638.699364] env[63197]: DEBUG nova.compute.utils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.700285] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.747s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.701893] env[63197]: INFO nova.compute.claims [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.710609] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Build of instance 2d8fb706-7400-43af-86fe-f90ac38393e0 was re-scheduled: Binding failed for port e6c7bf2d-771d-45c4-8e24-0ad3ae822ba5, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.711102] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.711336] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Acquiring lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.711477] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Acquired lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.711628] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.807208] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b79794-454a-0f15-6b5e-5c5783510be1, 'name': SearchDatastore_Task, 'duration_secs': 0.043003} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.807523] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.807858] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 638.808193] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.808193] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.808331] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 638.808526] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-432d787c-2a30-4c98-bb55-d19c62407b9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.816441] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 638.816621] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 638.817375] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a34ac90-6948-41d0-9453-b493f59851b3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.822213] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 638.822213] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d2afd0-9256-b413-2cf9-cf48d58fdf33" [ 638.822213] env[63197]: _type = "Task" [ 638.822213] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.829576] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d2afd0-9256-b413-2cf9-cf48d58fdf33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.228911] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.327403] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.333387] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d2afd0-9256-b413-2cf9-cf48d58fdf33, 'name': SearchDatastore_Task, 'duration_secs': 0.008098} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.334928] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cef94d52-2353-4495-aa41-a14fd46e135b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.341415] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 639.341415] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bc2888-99d3-f2e7-08c0-c96c1a4757bf" [ 639.341415] env[63197]: _type = "Task" [ 639.341415] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.350576] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bc2888-99d3-f2e7-08c0-c96c1a4757bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.837184] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Releasing lock "refresh_cache-2d8fb706-7400-43af-86fe-f90ac38393e0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.837184] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.837184] env[63197]: DEBUG nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.837598] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.852457] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bc2888-99d3-f2e7-08c0-c96c1a4757bf, 'name': SearchDatastore_Task, 'duration_secs': 0.011497} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.854776] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.855095] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 639.855712] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94e06cc7-46df-4b0a-b0ab-62215408836c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.857915] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.864098] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 639.864098] env[63197]: value = "task-1364064" [ 639.864098] env[63197]: _type = "Task" [ 639.864098] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.874094] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364064, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.127258] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31a5d89-e9f2-4352-b1e8-0f7d0cf048ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.139968] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc00454-9280-4576-918c-2bb9665c46c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.169561] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e81b537-0ab1-4773-beb4-b9078dcfb273 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.178638] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2320dcf0-7659-4d26-b6b9-b00db7e5cf27 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.195403] env[63197]: DEBUG nova.compute.provider_tree [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.361370] env[63197]: DEBUG nova.network.neutron [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.382964] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364064, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.699809] env[63197]: DEBUG nova.scheduler.client.report [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.868961] env[63197]: INFO nova.compute.manager [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] [instance: 2d8fb706-7400-43af-86fe-f90ac38393e0] Took 1.03 seconds to deallocate network for instance. [ 640.882230] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364064, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527647} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.882578] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 640.882752] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 640.884025] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc70d23e-cb4d-4577-bfae-1db86548ab5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.890879] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 640.890879] env[63197]: value = "task-1364065" [ 640.890879] env[63197]: _type = "Task" [ 640.890879] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.901416] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364065, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.209234] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.209809] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 641.213205] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.536s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.402824] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364065, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077749} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.403136] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 641.404398] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17619f61-d9d1-4b19-8c5e-7393458cbe3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.429819] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 641.431088] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-378cbeb3-f5eb-4512-89bb-ca8fe71fdb48 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.459295] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 641.459295] env[63197]: value = "task-1364067" [ 641.459295] env[63197]: _type = "Task" [ 641.459295] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.468272] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364067, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.721024] env[63197]: DEBUG nova.compute.utils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 641.723902] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 641.724224] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 641.796721] env[63197]: DEBUG nova.policy [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51f633ccfefa485ea49d334004d3f1a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7e03b368444f40128c3f4c5387a4d95e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 641.914450] env[63197]: INFO nova.scheduler.client.report [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Deleted allocations for instance 2d8fb706-7400-43af-86fe-f90ac38393e0 [ 641.970783] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364067, 'name': ReconfigVM_Task, 'duration_secs': 0.282622} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.973472] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Reconfigured VM instance instance-00000012 to attach disk [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 641.977119] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-837abbde-63cf-46d5-9285-f5099c8b8850 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.989512] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 641.989512] env[63197]: value = "task-1364068" [ 641.989512] env[63197]: _type = "Task" [ 641.989512] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.999615] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364068, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.207517] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Successfully created port: 44d43c99-1a55-4847-809d-e6abdc64cc3f {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 642.229104] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 642.254282] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f5b132-e272-492a-8ff2-9104ef932bff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.267806] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6005b86d-1b04-4550-a4a6-35c4c20f85cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.303034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274d980c-9e19-425d-8b76-a5a66b3fe956 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.312034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4966daef-3713-4045-9cd8-dfda83a3f9a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.327154] env[63197]: DEBUG nova.compute.provider_tree [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.432464] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cb95a75f-c841-47ad-9415-594ba5af7030 tempest-VolumesAssistedSnapshotsTest-386366189 tempest-VolumesAssistedSnapshotsTest-386366189-project-member] Lock "2d8fb706-7400-43af-86fe-f90ac38393e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.489s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.503568] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364068, 'name': Rename_Task, 'duration_secs': 0.142692} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.504234] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 642.504725] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc5d2a3f-b80a-450e-aaf6-80c75eb41b2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.513594] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 642.513594] env[63197]: value = "task-1364069" [ 642.513594] env[63197]: _type = "Task" [ 642.513594] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.524300] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364069, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.539320] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Acquiring lock "a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.539748] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Lock "a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.831342] env[63197]: DEBUG nova.scheduler.client.report [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.936951] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.026686] env[63197]: DEBUG oslo_vmware.api [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364069, 'name': PowerOnVM_Task, 'duration_secs': 0.489868} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.027135] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 643.027475] env[63197]: INFO nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Took 5.37 seconds to spawn the instance on the hypervisor. [ 643.028085] env[63197]: DEBUG nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 643.028962] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d076210-2903-4d76-ab3b-3160b3715aa3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.245255] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 643.285054] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 643.285054] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 643.285054] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 643.285232] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 643.285232] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 643.285388] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 643.286233] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 643.286834] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 643.286834] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 643.286834] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 643.287135] env[63197]: DEBUG nova.virt.hardware [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 643.288310] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d5880d-a778-45ec-a777-8e46b6b5968e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.302859] env[63197]: DEBUG nova.compute.manager [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Received event network-changed-44d43c99-1a55-4847-809d-e6abdc64cc3f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 643.302859] env[63197]: DEBUG nova.compute.manager [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Refreshing instance network info cache due to event network-changed-44d43c99-1a55-4847-809d-e6abdc64cc3f. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 643.302859] env[63197]: DEBUG oslo_concurrency.lockutils [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] Acquiring lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.302859] env[63197]: DEBUG oslo_concurrency.lockutils [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] Acquired lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.302859] env[63197]: DEBUG nova.network.neutron [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Refreshing network info cache for port 44d43c99-1a55-4847-809d-e6abdc64cc3f {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 643.312998] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0e85e1-7532-44cb-9713-4717a001f46e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.336353] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.123s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.337013] env[63197]: ERROR nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Traceback (most recent call last): [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self.driver.spawn(context, instance, image_meta, [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] vm_ref = self.build_virtual_machine(instance, [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.337013] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] for vif in network_info: [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return self._sync_wrapper(fn, *args, **kwargs) [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self.wait() [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self[:] = self._gt.wait() [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return self._exit_event.wait() [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] result = hub.switch() [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.337459] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return self.greenlet.switch() [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] result = function(*args, **kwargs) [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] return func(*args, **kwargs) [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] raise e [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] nwinfo = self.network_api.allocate_for_instance( [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] created_port_ids = self._update_ports_for_instance( [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] with excutils.save_and_reraise_exception(): [ 643.338022] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] self.force_reraise() [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] raise self.value [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] updated_port = self._update_port( [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] _ensure_no_port_binding_failure(port) [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] raise exception.PortBindingFailed(port_id=port['id']) [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] nova.exception.PortBindingFailed: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. [ 643.338477] env[63197]: ERROR nova.compute.manager [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] [ 643.338959] env[63197]: DEBUG nova.compute.utils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 643.338959] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.310s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.339170] env[63197]: DEBUG nova.objects.instance [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lazy-loading 'resources' on Instance uuid 99539d7f-cd93-4250-97cd-2449e03e517d {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 643.344060] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Build of instance 441988d7-b5c1-4996-a6fa-4a6e21f86b22 was re-scheduled: Binding failed for port a0c41c89-0eff-4df5-897a-3237572c810e, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 643.344635] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 643.345259] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Acquiring lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.345259] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Acquired lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.345259] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.458264] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.488218] env[63197]: ERROR nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. [ 643.488218] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 643.488218] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.488218] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 643.488218] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.488218] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 643.488218] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.488218] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 643.488218] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.488218] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 643.488218] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.488218] env[63197]: ERROR nova.compute.manager raise self.value [ 643.488218] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.488218] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 643.488218] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.488218] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 643.488678] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.488678] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 643.488678] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. [ 643.488678] env[63197]: ERROR nova.compute.manager [ 643.488678] env[63197]: Traceback (most recent call last): [ 643.488678] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 643.488678] env[63197]: listener.cb(fileno) [ 643.488678] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.488678] env[63197]: result = function(*args, **kwargs) [ 643.488678] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 643.488678] env[63197]: return func(*args, **kwargs) [ 643.488678] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.488678] env[63197]: raise e [ 643.488678] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.488678] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 643.488678] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.488678] env[63197]: created_port_ids = self._update_ports_for_instance( [ 643.488678] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.488678] env[63197]: with excutils.save_and_reraise_exception(): [ 643.488678] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.488678] env[63197]: self.force_reraise() [ 643.488678] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.488678] env[63197]: raise self.value [ 643.488678] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.488678] env[63197]: updated_port = self._update_port( [ 643.488678] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.488678] env[63197]: _ensure_no_port_binding_failure(port) [ 643.488678] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.488678] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 643.489528] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. [ 643.489528] env[63197]: Removing descriptor: 17 [ 643.489528] env[63197]: ERROR nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Traceback (most recent call last): [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] yield resources [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self.driver.spawn(context, instance, image_meta, [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.489528] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] vm_ref = self.build_virtual_machine(instance, [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] for vif in network_info: [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return self._sync_wrapper(fn, *args, **kwargs) [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self.wait() [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self[:] = self._gt.wait() [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return self._exit_event.wait() [ 643.489847] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] result = hub.switch() [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return self.greenlet.switch() [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] result = function(*args, **kwargs) [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return func(*args, **kwargs) [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] raise e [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] nwinfo = self.network_api.allocate_for_instance( [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 643.490366] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] created_port_ids = self._update_ports_for_instance( [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] with excutils.save_and_reraise_exception(): [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self.force_reraise() [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] raise self.value [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] updated_port = self._update_port( [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] _ensure_no_port_binding_failure(port) [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.490638] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] raise exception.PortBindingFailed(port_id=port['id']) [ 643.490892] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] nova.exception.PortBindingFailed: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. [ 643.490892] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] [ 643.490892] env[63197]: INFO nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Terminating instance [ 643.492248] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Acquiring lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.554909] env[63197]: INFO nova.compute.manager [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Took 39.16 seconds to build instance. [ 643.827947] env[63197]: DEBUG nova.network.neutron [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.868101] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.965414] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.971533] env[63197]: DEBUG nova.network.neutron [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.015398] env[63197]: INFO nova.compute.manager [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Rebuilding instance [ 644.056490] env[63197]: DEBUG oslo_concurrency.lockutils [None req-850ded16-a560-4dc1-b43c-301e9ae804fe tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.128s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.062035] env[63197]: DEBUG nova.compute.manager [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 644.062760] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c2670c-4a04-41b8-9824-56d39d0a86ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.212316] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e228f1c8-e27d-46a9-9620-16c8f37202a1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.220369] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e67004-9abe-4aa0-a4c9-a18f428ce78d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.251032] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3597e5-515b-49f1-975f-d7b39fff8e3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.258885] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffebe1c-b162-4084-8c10-c6c524f50cf6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.272632] env[63197]: DEBUG nova.compute.provider_tree [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.470505] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Releasing lock "refresh_cache-441988d7-b5c1-4996-a6fa-4a6e21f86b22" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.470963] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 644.471259] env[63197]: DEBUG nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.471491] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.473451] env[63197]: DEBUG oslo_concurrency.lockutils [req-7771a785-b277-45d9-8cb9-c0aa09cacdd7 req-100b8fe0-ac63-491b-b71a-0a43e8416672 service nova] Releasing lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.473963] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Acquired lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.474920] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 644.502791] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.565887] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 644.575853] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 644.576145] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbca6d6a-b6af-4478-8df7-1abbd5612082 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.586069] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 644.586069] env[63197]: value = "task-1364071" [ 644.586069] env[63197]: _type = "Task" [ 644.586069] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.598850] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364071, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.776625] env[63197]: DEBUG nova.scheduler.client.report [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.002150] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.005619] env[63197]: DEBUG nova.network.neutron [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.091226] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.098100] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364071, 'name': PowerOffVM_Task, 'duration_secs': 0.170576} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.098365] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 645.098572] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 645.099313] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e596bb-9692-4bb2-990b-7b35af1d0adb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.106571] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 645.106798] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81dfe12c-ac64-4784-89be-f026db3a2ca8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.134353] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 645.134353] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 645.134624] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Deleting the datastore file [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 645.134671] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20aa9b1c-dc56-4eb4-aee3-0408ce395561 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.142728] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 645.142728] env[63197]: value = "task-1364073" [ 645.142728] env[63197]: _type = "Task" [ 645.142728] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.151858] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364073, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.157664] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.281780] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.284145] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.839s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.286207] env[63197]: INFO nova.compute.claims [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.302871] env[63197]: INFO nova.scheduler.client.report [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Deleted allocations for instance 99539d7f-cd93-4250-97cd-2449e03e517d [ 645.352829] env[63197]: DEBUG nova.compute.manager [req-12231ec2-b808-4918-a821-f2506585e528 req-8bfd4d17-fcdf-45ae-aefa-21fc036d1801 service nova] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Received event network-vif-deleted-44d43c99-1a55-4847-809d-e6abdc64cc3f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 645.512995] env[63197]: INFO nova.compute.manager [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] [instance: 441988d7-b5c1-4996-a6fa-4a6e21f86b22] Took 1.04 seconds to deallocate network for instance. [ 645.653635] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364073, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101562} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.653922] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 645.654122] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 645.654345] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.659871] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Releasing lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.661416] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.661416] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 645.661416] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3421642-31c4-4f9c-a254-c694acc2c395 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.672330] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-693d9737-0acc-4cb9-962c-2762c9026312 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.694988] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f01a025d-2cb9-4cb9-878a-b05e318c21b9 could not be found. [ 645.695228] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.695410] env[63197]: INFO nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 645.695654] env[63197]: DEBUG oslo.service.loopingcall [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.696896] env[63197]: DEBUG nova.compute.manager [-] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.696896] env[63197]: DEBUG nova.network.neutron [-] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 645.717985] env[63197]: DEBUG nova.network.neutron [-] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.809395] env[63197]: DEBUG oslo_concurrency.lockutils [None req-77194ed6-137d-4eac-8045-8245042940c2 tempest-ServersAdmin275Test-2086138891 tempest-ServersAdmin275Test-2086138891-project-member] Lock "99539d7f-cd93-4250-97cd-2449e03e517d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.536s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.223471] env[63197]: DEBUG nova.network.neutron [-] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.549268] env[63197]: INFO nova.scheduler.client.report [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Deleted allocations for instance 441988d7-b5c1-4996-a6fa-4a6e21f86b22 [ 646.700614] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 646.700859] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 646.701025] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 646.701392] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 646.701392] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 646.701470] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 646.701673] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 646.701861] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 646.702017] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 646.702210] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 646.702383] env[63197]: DEBUG nova.virt.hardware [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 646.703263] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f903dc-1ce6-44e5-9d25-66c38ec38670 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.712923] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb43ca15-31f7-418a-9449-c98e5bcb4981 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.720644] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb17b52a-009a-46bf-82c8-b6bd81815042 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.735805] env[63197]: INFO nova.compute.manager [-] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Took 1.04 seconds to deallocate network for instance. [ 646.736316] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 646.742663] env[63197]: DEBUG oslo.service.loopingcall [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.744987] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 646.746179] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554cb784-8e8e-43a5-8bc3-23a594c5648e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.750729] env[63197]: DEBUG nova.compute.claims [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 646.750941] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.751171] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e0db208-99c4-46fd-8da0-4effb586451d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.793979] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d86edd-acd1-4b0f-aa36-0f85d3b7c64b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.796633] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 646.796633] env[63197]: value = "task-1364074" [ 646.796633] env[63197]: _type = "Task" [ 646.796633] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.803888] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6677c5-914b-4339-99da-a8f6673db264 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.813083] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364074, 'name': CreateVM_Task} progress is 15%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.823231] env[63197]: DEBUG nova.compute.provider_tree [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.063134] env[63197]: DEBUG oslo_concurrency.lockutils [None req-45089922-aa48-4726-a20d-8643fc9c77b9 tempest-ServersWithSpecificFlavorTestJSON-1115321942 tempest-ServersWithSpecificFlavorTestJSON-1115321942-project-member] Lock "441988d7-b5c1-4996-a6fa-4a6e21f86b22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.577s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.308419] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364074, 'name': CreateVM_Task, 'duration_secs': 0.314501} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.308900] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 647.309496] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.309664] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.310086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 647.310365] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e39b83e-a7de-473f-bf71-80ee55f4a817 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.316054] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 647.316054] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fd464e-868a-2d7f-50a9-dd4d1f5c252e" [ 647.316054] env[63197]: _type = "Task" [ 647.316054] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.325089] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fd464e-868a-2d7f-50a9-dd4d1f5c252e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.326924] env[63197]: DEBUG nova.scheduler.client.report [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.543825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Acquiring lock "e28f14fc-dd09-4baf-9be9-cba280f47f80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.543825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Lock "e28f14fc-dd09-4baf-9be9-cba280f47f80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.568052] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 647.828573] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fd464e-868a-2d7f-50a9-dd4d1f5c252e, 'name': SearchDatastore_Task, 'duration_secs': 0.009732} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.828573] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.828573] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 647.828843] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.828843] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.828914] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 647.829185] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06c5da2e-17eb-4c97-9971-05b0699e79df {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.832056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.832056] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.834932] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.281s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.836426] env[63197]: INFO nova.compute.claims [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.850697] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 647.851695] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 647.851695] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c17fc48-b12b-42ad-971f-68591a6169c2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.859368] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 647.859368] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ee38cc-4728-be02-ea4d-10128abfccc8" [ 647.859368] env[63197]: _type = "Task" [ 647.859368] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.869139] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ee38cc-4728-be02-ea4d-10128abfccc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.098406] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.337153] env[63197]: DEBUG nova.compute.utils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.338991] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.339321] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.376637] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ee38cc-4728-be02-ea4d-10128abfccc8, 'name': SearchDatastore_Task, 'duration_secs': 0.015401} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.380248] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd16877a-3d4b-4385-ab9c-4c704e7df76e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.383498] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 648.383498] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ea6c6f-07b4-9d0a-8512-c98cb2d7636a" [ 648.383498] env[63197]: _type = "Task" [ 648.383498] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.394375] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ea6c6f-07b4-9d0a-8512-c98cb2d7636a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.411425] env[63197]: DEBUG nova.policy [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cd1ecc7212045eeaab6358edace72ed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '14190e3f6975452dbfc77ab3a1f51094', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.845149] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.897129] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ea6c6f-07b4-9d0a-8512-c98cb2d7636a, 'name': SearchDatastore_Task, 'duration_secs': 0.015118} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.900055] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.900450] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 648.901264] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9bf79709-66d8-4703-88d5-fd7a20bd5581 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.913436] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 648.913436] env[63197]: value = "task-1364075" [ 648.913436] env[63197]: _type = "Task" [ 648.913436] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.926878] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.013306] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Successfully created port: 9223f478-9a65-4de3-ad95-dcd4a96500c4 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.229960] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85376815-988f-43a2-b21d-8944d3de0531 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.239369] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f815dc-34c0-4c59-a8f0-be5dca59419c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.274705] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c82dac-2f4f-4dc0-887c-ccdfa8fd34d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.287665] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97213c93-5986-4626-a4c0-9ec04b68adbf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.305908] env[63197]: DEBUG nova.compute.provider_tree [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.429632] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364075, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.810246] env[63197]: DEBUG nova.scheduler.client.report [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.854820] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.887956] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.888364] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.888698] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.888698] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.888868] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.889017] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.889474] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.889692] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.889925] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.890143] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.890373] env[63197]: DEBUG nova.virt.hardware [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.891500] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61891938-2fdc-439e-9e11-ed48f30fa99f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.900739] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983496e1-a552-40a9-8d27-5b7182d2b090 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.925416] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.723374} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.926010] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 649.926294] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 649.926572] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fd25cf6-5577-4057-b3d6-a102fdc450a2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.935255] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 649.935255] env[63197]: value = "task-1364076" [ 649.935255] env[63197]: _type = "Task" [ 649.935255] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.948040] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364076, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.316948] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.317711] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.321417] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.096s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.447413] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364076, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078811} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.447860] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 650.448509] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7c361d-1964-4956-8acd-9dfd03e156ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.469819] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 650.470142] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef1ea1b0-e906-4256-859d-057621fb8106 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.492811] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 650.492811] env[63197]: value = "task-1364077" [ 650.492811] env[63197]: _type = "Task" [ 650.492811] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.502551] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364077, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.825287] env[63197]: DEBUG nova.compute.utils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.830138] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.830138] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 651.002219] env[63197]: DEBUG nova.policy [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a2d841977ff4233951e3ed95c5d91ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e61a2b00ac24d55b4bbba8977f9704c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.003410] env[63197]: DEBUG nova.compute.manager [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Received event network-changed-9223f478-9a65-4de3-ad95-dcd4a96500c4 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.003599] env[63197]: DEBUG nova.compute.manager [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Refreshing instance network info cache due to event network-changed-9223f478-9a65-4de3-ad95-dcd4a96500c4. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 651.003872] env[63197]: DEBUG oslo_concurrency.lockutils [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] Acquiring lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.003947] env[63197]: DEBUG oslo_concurrency.lockutils [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] Acquired lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.004249] env[63197]: DEBUG nova.network.neutron [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Refreshing network info cache for port 9223f478-9a65-4de3-ad95-dcd4a96500c4 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 651.025512] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364077, 'name': ReconfigVM_Task, 'duration_secs': 0.299218} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.025808] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Reconfigured VM instance instance-00000012 to attach disk [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f/a066e6e1-12a4-455b-87cc-e3d6a6a17c1f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 651.026945] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2a31dba-9596-4860-ac92-d8bb0ebae109 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.037774] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 651.037774] env[63197]: value = "task-1364078" [ 651.037774] env[63197]: _type = "Task" [ 651.037774] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.060825] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364078, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.074128] env[63197]: DEBUG nova.network.neutron [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.199116] env[63197]: ERROR nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. [ 651.199116] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.199116] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.199116] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.199116] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.199116] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.199116] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.199116] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.199116] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.199116] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 651.199116] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.199116] env[63197]: ERROR nova.compute.manager raise self.value [ 651.199116] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.199116] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.199116] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.199116] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.199593] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.199593] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.199593] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. [ 651.199593] env[63197]: ERROR nova.compute.manager [ 651.199593] env[63197]: Traceback (most recent call last): [ 651.199593] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.199593] env[63197]: listener.cb(fileno) [ 651.199593] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.199593] env[63197]: result = function(*args, **kwargs) [ 651.199593] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.199593] env[63197]: return func(*args, **kwargs) [ 651.199593] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.199593] env[63197]: raise e [ 651.199593] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.199593] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 651.199593] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.199593] env[63197]: created_port_ids = self._update_ports_for_instance( [ 651.199593] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.199593] env[63197]: with excutils.save_and_reraise_exception(): [ 651.199593] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.199593] env[63197]: self.force_reraise() [ 651.199593] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.199593] env[63197]: raise self.value [ 651.199593] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.199593] env[63197]: updated_port = self._update_port( [ 651.199593] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.199593] env[63197]: _ensure_no_port_binding_failure(port) [ 651.199593] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.199593] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.200429] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. [ 651.200429] env[63197]: Removing descriptor: 17 [ 651.206598] env[63197]: ERROR nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Traceback (most recent call last): [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] yield resources [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self.driver.spawn(context, instance, image_meta, [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] vm_ref = self.build_virtual_machine(instance, [ 651.206598] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] for vif in network_info: [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return self._sync_wrapper(fn, *args, **kwargs) [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self.wait() [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self[:] = self._gt.wait() [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return self._exit_event.wait() [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.206857] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] result = hub.switch() [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return self.greenlet.switch() [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] result = function(*args, **kwargs) [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return func(*args, **kwargs) [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] raise e [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] nwinfo = self.network_api.allocate_for_instance( [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] created_port_ids = self._update_ports_for_instance( [ 651.207159] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] with excutils.save_and_reraise_exception(): [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self.force_reraise() [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] raise self.value [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] updated_port = self._update_port( [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] _ensure_no_port_binding_failure(port) [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] raise exception.PortBindingFailed(port_id=port['id']) [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] nova.exception.PortBindingFailed: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. [ 651.207440] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] [ 651.207735] env[63197]: INFO nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Terminating instance [ 651.207735] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Acquiring lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.219513] env[63197]: DEBUG nova.network.neutron [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.238319] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb295b93-67be-493b-8a00-b12365026be1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.247060] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38bdf000-89cd-4146-a344-bf6d47a837e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.282676] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e878c49-7344-4470-9a17-4ab8cc5f1a78 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.291163] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7103229-9a54-4499-a27e-126559776dbf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.306408] env[63197]: DEBUG nova.compute.provider_tree [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.332193] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.556321] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364078, 'name': Rename_Task, 'duration_secs': 0.175549} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.556529] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 651.556888] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ec51f3d-8830-4203-a681-ce2aa6281d23 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.568355] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 651.568355] env[63197]: value = "task-1364079" [ 651.568355] env[63197]: _type = "Task" [ 651.568355] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.580063] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.724421] env[63197]: DEBUG oslo_concurrency.lockutils [req-73ecc2c9-48e3-47d1-ac21-f62ddc7c2635 req-9fab24b7-ffac-48fb-a551-d9d4cde6afdb service nova] Releasing lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.725029] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Acquired lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.725176] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.813021] env[63197]: DEBUG nova.scheduler.client.report [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.924875] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Successfully created port: d4c6030d-61e8-454b-8aea-337e0cf65ba2 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.080987] env[63197]: DEBUG oslo_vmware.api [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364079, 'name': PowerOnVM_Task, 'duration_secs': 0.451163} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.081291] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 652.081489] env[63197]: DEBUG nova.compute.manager [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 652.082564] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95265e82-3a1c-4bbe-b68a-c5aa3e407a66 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.262141] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.319396] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.320158] env[63197]: ERROR nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] Traceback (most recent call last): [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self.driver.spawn(context, instance, image_meta, [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] vm_ref = self.build_virtual_machine(instance, [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.320158] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] for vif in network_info: [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return self._sync_wrapper(fn, *args, **kwargs) [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self.wait() [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self[:] = self._gt.wait() [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return self._exit_event.wait() [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] result = hub.switch() [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.320466] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return self.greenlet.switch() [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] result = function(*args, **kwargs) [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] return func(*args, **kwargs) [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] raise e [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] nwinfo = self.network_api.allocate_for_instance( [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] created_port_ids = self._update_ports_for_instance( [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] with excutils.save_and_reraise_exception(): [ 652.320769] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] self.force_reraise() [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] raise self.value [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] updated_port = self._update_port( [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] _ensure_no_port_binding_failure(port) [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] raise exception.PortBindingFailed(port_id=port['id']) [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] nova.exception.PortBindingFailed: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. [ 652.321105] env[63197]: ERROR nova.compute.manager [instance: bb7f3278-8678-4053-b957-9ceee430089d] [ 652.321442] env[63197]: DEBUG nova.compute.utils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.322784] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.571s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.326117] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Build of instance bb7f3278-8678-4053-b957-9ceee430089d was re-scheduled: Binding failed for port 578d90c4-64f2-4194-9d29-f171e09a6c41, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 652.330467] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 652.331292] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.331292] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquired lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.331292] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.347673] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.384909] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.385774] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.385991] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.386237] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.386418] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.386596] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.386837] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.387040] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.387252] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.387447] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.387667] env[63197]: DEBUG nova.virt.hardware [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.388606] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-866f2284-329f-434b-af33-9f4fa60cd58c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.397772] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e10a125-4cf5-4e06-b4f1-4a67faaf9232 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.458859] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.611054] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.874139] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.964035] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Releasing lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.964035] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.964035] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.964035] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-481a2582-5f69-41a6-a04c-0051ba15d1e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.974854] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff7d2fa-78d2-4333-8754-7b25e81e702d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.010551] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b52f3d4b-7699-484d-8b4e-84393848fb37 could not be found. [ 653.011302] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.011302] env[63197]: INFO nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Took 0.05 seconds to destroy the instance on the hypervisor. [ 653.011302] env[63197]: DEBUG oslo.service.loopingcall [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.011472] env[63197]: DEBUG nova.compute.manager [-] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.011505] env[63197]: DEBUG nova.network.neutron [-] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.054602] env[63197]: DEBUG nova.network.neutron [-] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.066318] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.078702] env[63197]: DEBUG nova.compute.manager [req-97cc7de2-a45f-4f81-bff5-d9d708d01f96 req-6b511c34-8348-4a4f-b502-f61ad2e72abf service nova] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Received event network-vif-deleted-9223f478-9a65-4de3-ad95-dcd4a96500c4 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.245013] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.245341] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.245567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.245755] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.245956] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.252467] env[63197]: INFO nova.compute.manager [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Terminating instance [ 653.255405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "refresh_cache-a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.255405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquired lock "refresh_cache-a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.255405] env[63197]: DEBUG nova.network.neutron [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.345995] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c880a02-582f-47e7-8876-d92781639c0a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.356427] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf7e64b-8e35-4272-a708-931de921217c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.395605] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30547210-a3a4-4972-a984-acdcdf7fb15c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.404277] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ad1f19-1236-4ee4-b5b4-92ff06338104 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.420977] env[63197]: DEBUG nova.compute.provider_tree [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.540146] env[63197]: DEBUG nova.compute.manager [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Received event network-changed-d4c6030d-61e8-454b-8aea-337e0cf65ba2 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.540146] env[63197]: DEBUG nova.compute.manager [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Refreshing instance network info cache due to event network-changed-d4c6030d-61e8-454b-8aea-337e0cf65ba2. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 653.540146] env[63197]: DEBUG oslo_concurrency.lockutils [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] Acquiring lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.540146] env[63197]: DEBUG oslo_concurrency.lockutils [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] Acquired lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.540146] env[63197]: DEBUG nova.network.neutron [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Refreshing network info cache for port d4c6030d-61e8-454b-8aea-337e0cf65ba2 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 653.559530] env[63197]: DEBUG nova.network.neutron [-] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.571960] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Releasing lock "refresh_cache-bb7f3278-8678-4053-b957-9ceee430089d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.571960] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 653.571960] env[63197]: DEBUG nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.571960] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.595139] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.642010] env[63197]: ERROR nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. [ 653.642010] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.642010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.642010] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.642010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.642010] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.642010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.642010] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.642010] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.642010] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 653.642010] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.642010] env[63197]: ERROR nova.compute.manager raise self.value [ 653.642010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.642010] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.642010] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.642010] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.642706] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.642706] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.642706] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. [ 653.642706] env[63197]: ERROR nova.compute.manager [ 653.642706] env[63197]: Traceback (most recent call last): [ 653.642706] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.642706] env[63197]: listener.cb(fileno) [ 653.642706] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.642706] env[63197]: result = function(*args, **kwargs) [ 653.642706] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.642706] env[63197]: return func(*args, **kwargs) [ 653.642706] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.642706] env[63197]: raise e [ 653.642706] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.642706] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 653.642706] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.642706] env[63197]: created_port_ids = self._update_ports_for_instance( [ 653.642706] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.642706] env[63197]: with excutils.save_and_reraise_exception(): [ 653.642706] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.642706] env[63197]: self.force_reraise() [ 653.642706] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.642706] env[63197]: raise self.value [ 653.642706] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.642706] env[63197]: updated_port = self._update_port( [ 653.642706] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.642706] env[63197]: _ensure_no_port_binding_failure(port) [ 653.642706] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.642706] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.643476] env[63197]: nova.exception.PortBindingFailed: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. [ 653.643476] env[63197]: Removing descriptor: 19 [ 653.643476] env[63197]: ERROR nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Traceback (most recent call last): [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] yield resources [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self.driver.spawn(context, instance, image_meta, [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.643476] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] vm_ref = self.build_virtual_machine(instance, [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] for vif in network_info: [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return self._sync_wrapper(fn, *args, **kwargs) [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self.wait() [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self[:] = self._gt.wait() [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return self._exit_event.wait() [ 653.643751] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] result = hub.switch() [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return self.greenlet.switch() [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] result = function(*args, **kwargs) [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return func(*args, **kwargs) [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] raise e [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] nwinfo = self.network_api.allocate_for_instance( [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.644083] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] created_port_ids = self._update_ports_for_instance( [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] with excutils.save_and_reraise_exception(): [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self.force_reraise() [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] raise self.value [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] updated_port = self._update_port( [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] _ensure_no_port_binding_failure(port) [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.644501] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] raise exception.PortBindingFailed(port_id=port['id']) [ 653.644782] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] nova.exception.PortBindingFailed: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. [ 653.644782] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] [ 653.644782] env[63197]: INFO nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Terminating instance [ 653.645281] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Acquiring lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.781695] env[63197]: DEBUG nova.network.neutron [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.843575] env[63197]: DEBUG nova.network.neutron [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.924016] env[63197]: DEBUG nova.scheduler.client.report [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.061153] env[63197]: DEBUG nova.network.neutron [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.063676] env[63197]: INFO nova.compute.manager [-] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Took 1.05 seconds to deallocate network for instance. [ 654.066276] env[63197]: DEBUG nova.compute.claims [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.066451] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.100178] env[63197]: DEBUG nova.network.neutron [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.156056] env[63197]: DEBUG nova.network.neutron [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.351053] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Releasing lock "refresh_cache-a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.352394] env[63197]: DEBUG nova.compute.manager [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 654.352394] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 654.353252] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427fa2ab-c410-4ffe-a5ff-b19214bd8717 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.362626] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 654.362904] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8fa5bc3-d398-4b10-9c6d-1ab0bfee2832 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.370272] env[63197]: DEBUG oslo_vmware.api [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 654.370272] env[63197]: value = "task-1364080" [ 654.370272] env[63197]: _type = "Task" [ 654.370272] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.378979] env[63197]: DEBUG oslo_vmware.api [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364080, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.428478] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "6c3137a4-4e51-4bf9-b7d3-779de61b69b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.428747] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "6c3137a4-4e51-4bf9-b7d3-779de61b69b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.433352] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.110s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.433668] env[63197]: ERROR nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Traceback (most recent call last): [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self.driver.spawn(context, instance, image_meta, [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] vm_ref = self.build_virtual_machine(instance, [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.433668] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] for vif in network_info: [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return self._sync_wrapper(fn, *args, **kwargs) [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self.wait() [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self[:] = self._gt.wait() [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return self._exit_event.wait() [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] result = hub.switch() [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.433995] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return self.greenlet.switch() [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] result = function(*args, **kwargs) [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] return func(*args, **kwargs) [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] raise e [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] nwinfo = self.network_api.allocate_for_instance( [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] created_port_ids = self._update_ports_for_instance( [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] with excutils.save_and_reraise_exception(): [ 654.434326] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] self.force_reraise() [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] raise self.value [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] updated_port = self._update_port( [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] _ensure_no_port_binding_failure(port) [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] raise exception.PortBindingFailed(port_id=port['id']) [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] nova.exception.PortBindingFailed: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. [ 654.434772] env[63197]: ERROR nova.compute.manager [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] [ 654.435283] env[63197]: DEBUG nova.compute.utils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.435683] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.073s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.439020] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Build of instance 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282 was re-scheduled: Binding failed for port e8b871f2-4263-48c9-9202-655d88ff6865, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 654.440114] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 654.440114] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Acquiring lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.440114] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Acquired lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.440114] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.472492] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "15120438-f2a8-4295-a70f-6ec8afdd9030" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.472803] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "15120438-f2a8-4295-a70f-6ec8afdd9030" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.602612] env[63197]: INFO nova.compute.manager [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: bb7f3278-8678-4053-b957-9ceee430089d] Took 1.03 seconds to deallocate network for instance. [ 654.658891] env[63197]: DEBUG oslo_concurrency.lockutils [req-03be769b-2eda-474a-9609-b0940dfac493 req-c902dc27-425a-4b1d-99fa-870dfc012e2e service nova] Releasing lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.659633] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Acquired lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.659858] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.883801] env[63197]: DEBUG oslo_vmware.api [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364080, 'name': PowerOffVM_Task, 'duration_secs': 0.134451} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.884173] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 654.884353] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 654.884611] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5958cb4-4701-4b17-9099-df3e41f481c4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.913161] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 654.913409] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 654.913591] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Deleting the datastore file [datastore2] a066e6e1-12a4-455b-87cc-e3d6a6a17c1f {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 654.913861] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb40c389-0f28-4faf-835f-1b8ec7a1585a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.921690] env[63197]: DEBUG oslo_vmware.api [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for the task: (returnval){ [ 654.921690] env[63197]: value = "task-1364082" [ 654.921690] env[63197]: _type = "Task" [ 654.921690] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.931993] env[63197]: DEBUG oslo_vmware.api [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364082, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.968568] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.181349] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.281792] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.340008] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72363509-c53c-4167-9a19-f94b4b7fa6b7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.345699] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.350407] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8370f84-c674-48b0-97d3-21239a3f73c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.384490] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45996c22-26fb-4819-9224-f82d6acea480 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.393555] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd372e0-1e08-4df9-97fd-6e3d8b5a7b27 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.408148] env[63197]: DEBUG nova.compute.provider_tree [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.435511] env[63197]: DEBUG oslo_vmware.api [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Task: {'id': task-1364082, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100352} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.435511] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 655.435511] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 655.435511] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.435511] env[63197]: INFO nova.compute.manager [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Took 1.08 seconds to destroy the instance on the hypervisor. [ 655.435905] env[63197]: DEBUG oslo.service.loopingcall [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.435905] env[63197]: DEBUG nova.compute.manager [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.435905] env[63197]: DEBUG nova.network.neutron [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.453305] env[63197]: DEBUG nova.network.neutron [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.565851] env[63197]: DEBUG nova.compute.manager [req-9d9d498f-1cea-4134-a5e6-05353dba2820 req-bcc21a79-c598-42cf-859a-f57ceddcff94 service nova] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Received event network-vif-deleted-d4c6030d-61e8-454b-8aea-337e0cf65ba2 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.639075] env[63197]: INFO nova.scheduler.client.report [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Deleted allocations for instance bb7f3278-8678-4053-b957-9ceee430089d [ 655.784390] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Releasing lock "refresh_cache-5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.784714] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 655.784823] env[63197]: DEBUG nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.784998] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.811186] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.851836] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Releasing lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.851836] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.851836] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.851836] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48f8ee48-6fe8-4489-b361-0ae025fe572c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.864124] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72e8386-d55e-4b80-ab2d-8b9841dfeb3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.889446] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a18802fb-2a73-4bb5-aab9-5a7eb599a5a6 could not be found. [ 655.889804] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.890130] env[63197]: INFO nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.890505] env[63197]: DEBUG oslo.service.loopingcall [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.890853] env[63197]: DEBUG nova.compute.manager [-] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.891077] env[63197]: DEBUG nova.network.neutron [-] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.913842] env[63197]: DEBUG nova.scheduler.client.report [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.921080] env[63197]: DEBUG nova.network.neutron [-] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.956314] env[63197]: DEBUG nova.network.neutron [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.148206] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fffe53f7-1eba-4d9c-8c61-4e646e926cfb tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "bb7f3278-8678-4053-b957-9ceee430089d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.526s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.310584] env[63197]: DEBUG nova.network.neutron [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.421015] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.421015] env[63197]: ERROR nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. [ 656.421015] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Traceback (most recent call last): [ 656.421015] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.421015] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self.driver.spawn(context, instance, image_meta, [ 656.421015] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 656.421015] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.421015] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.421015] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] vm_ref = self.build_virtual_machine(instance, [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] for vif in network_info: [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] return self._sync_wrapper(fn, *args, **kwargs) [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self.wait() [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self[:] = self._gt.wait() [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] return self._exit_event.wait() [ 656.421608] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] current.throw(*self._exc) [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] result = function(*args, **kwargs) [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] return func(*args, **kwargs) [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] raise e [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] nwinfo = self.network_api.allocate_for_instance( [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] created_port_ids = self._update_ports_for_instance( [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 656.421923] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] with excutils.save_and_reraise_exception(): [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] self.force_reraise() [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] raise self.value [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] updated_port = self._update_port( [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] _ensure_no_port_binding_failure(port) [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] raise exception.PortBindingFailed(port_id=port['id']) [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] nova.exception.PortBindingFailed: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. [ 656.422256] env[63197]: ERROR nova.compute.manager [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] [ 656.422524] env[63197]: DEBUG nova.compute.utils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.426585] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Build of instance d84f11a3-ed86-4fc4-9092-0babc7fccb3f was re-scheduled: Binding failed for port 8bee1fbe-7a99-4330-af05-7834adb33c6e, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.428690] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.428898] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquiring lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.429062] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Acquired lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.429276] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 656.430969] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.572s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.431972] env[63197]: INFO nova.compute.claims [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.435893] env[63197]: DEBUG nova.network.neutron [-] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.462038] env[63197]: INFO nova.compute.manager [-] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Took 1.03 seconds to deallocate network for instance. [ 656.651072] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 656.813894] env[63197]: INFO nova.compute.manager [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] [instance: 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282] Took 1.03 seconds to deallocate network for instance. [ 656.946248] env[63197]: INFO nova.compute.manager [-] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Took 1.06 seconds to deallocate network for instance. [ 656.954009] env[63197]: DEBUG nova.compute.claims [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.954568] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.966713] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.970448] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.139937] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.182193] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.642733] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Releasing lock "refresh_cache-d84f11a3-ed86-4fc4-9092-0babc7fccb3f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.643373] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.643373] env[63197]: DEBUG nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.643478] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 657.661846] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.848527] env[63197]: INFO nova.scheduler.client.report [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Deleted allocations for instance 5ceaa99b-2d75-42f4-a49a-4fd1e56d0282 [ 657.905774] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18281121-6f10-40ca-a6f3-1ecf19c0e36e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.915752] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f68030e-e692-418c-ab00-09fc4a50f6f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.950802] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c720094-63fe-4eff-90c7-07831e4198f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.959498] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f0060c-6da0-4686-8d89-d59db1457735 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.975390] env[63197]: DEBUG nova.compute.provider_tree [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.164060] env[63197]: DEBUG nova.network.neutron [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.361616] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b02393a8-ea81-45c7-9b06-9496792a2042 tempest-ServersV294TestFqdnHostnames-648359238 tempest-ServersV294TestFqdnHostnames-648359238-project-member] Lock "5ceaa99b-2d75-42f4-a49a-4fd1e56d0282" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.031s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.478861] env[63197]: DEBUG nova.scheduler.client.report [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.666602] env[63197]: INFO nova.compute.manager [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] [instance: d84f11a3-ed86-4fc4-9092-0babc7fccb3f] Took 1.02 seconds to deallocate network for instance. [ 658.865145] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 658.988988] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.988988] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 658.996862] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.532s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.996862] env[63197]: INFO nova.compute.claims [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.390319] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.507735] env[63197]: DEBUG nova.compute.utils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.509470] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.509641] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 659.575098] env[63197]: DEBUG nova.policy [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a482998aa094b07930690ee4c0f52a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c7edb4c5c5448a785e14223d1c008dc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 659.709055] env[63197]: INFO nova.scheduler.client.report [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Deleted allocations for instance d84f11a3-ed86-4fc4-9092-0babc7fccb3f [ 659.785440] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "b5935562-9a77-4b92-ac73-a7af496ccced" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.785760] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "b5935562-9a77-4b92-ac73-a7af496ccced" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.012817] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.098168] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Successfully created port: 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.227581] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6b3fb6d6-f82b-48b6-a875-fa857a5d5c5e tempest-ServersAdminTestJSON-1506909068 tempest-ServersAdminTestJSON-1506909068-project-member] Lock "d84f11a3-ed86-4fc4-9092-0babc7fccb3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.909s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.434509] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b6c9a3-2fc5-4095-b751-a01db9768247 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.444297] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e08448-9ee4-4cd4-bce2-f48c9134d7f3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.484307] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeff62bd-063d-418a-a5a2-8b5ed59eed18 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.494950] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f30bf9-4707-4179-8c9f-6225810282fe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.509840] env[63197]: DEBUG nova.compute.provider_tree [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.730906] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.013267] env[63197]: DEBUG nova.scheduler.client.report [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.029077] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.071012] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.071373] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.071570] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.072310] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.072310] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.072310] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.072310] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.072492] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.072627] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.072790] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.072955] env[63197]: DEBUG nova.virt.hardware [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.073821] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbeb0f55-b759-46be-a738-49ff5b022e14 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.083711] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca449cb8-d9c6-4a06-9d08-436a41453cd4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.260440] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.518639] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.519249] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.529718] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.433s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.529718] env[63197]: INFO nova.compute.claims [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.791317] env[63197]: DEBUG nova.compute.manager [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Received event network-changed-3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.794111] env[63197]: DEBUG nova.compute.manager [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Refreshing instance network info cache due to event network-changed-3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 661.794111] env[63197]: DEBUG oslo_concurrency.lockutils [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] Acquiring lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.794111] env[63197]: DEBUG oslo_concurrency.lockutils [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] Acquired lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.794111] env[63197]: DEBUG nova.network.neutron [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Refreshing network info cache for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 661.839334] env[63197]: ERROR nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. [ 661.839334] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.839334] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.839334] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.839334] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.839334] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.839334] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.839334] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.839334] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.839334] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 661.839334] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.839334] env[63197]: ERROR nova.compute.manager raise self.value [ 661.839334] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.839334] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.839334] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.839334] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.839769] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.839769] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.839769] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. [ 661.839769] env[63197]: ERROR nova.compute.manager [ 661.842018] env[63197]: Traceback (most recent call last): [ 661.842018] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.842018] env[63197]: listener.cb(fileno) [ 661.842018] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.842018] env[63197]: result = function(*args, **kwargs) [ 661.842018] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 661.842018] env[63197]: return func(*args, **kwargs) [ 661.842018] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.842018] env[63197]: raise e [ 661.842018] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.842018] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 661.842018] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.842018] env[63197]: created_port_ids = self._update_ports_for_instance( [ 661.842018] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.842018] env[63197]: with excutils.save_and_reraise_exception(): [ 661.842018] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.842018] env[63197]: self.force_reraise() [ 661.842018] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.842018] env[63197]: raise self.value [ 661.842018] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.842018] env[63197]: updated_port = self._update_port( [ 661.842018] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.842018] env[63197]: _ensure_no_port_binding_failure(port) [ 661.842018] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.842018] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.842018] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. [ 661.842018] env[63197]: Removing descriptor: 18 [ 661.842761] env[63197]: ERROR nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Traceback (most recent call last): [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] yield resources [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self.driver.spawn(context, instance, image_meta, [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] vm_ref = self.build_virtual_machine(instance, [ 661.842761] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] for vif in network_info: [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return self._sync_wrapper(fn, *args, **kwargs) [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self.wait() [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self[:] = self._gt.wait() [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return self._exit_event.wait() [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.843048] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] result = hub.switch() [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return self.greenlet.switch() [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] result = function(*args, **kwargs) [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return func(*args, **kwargs) [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] raise e [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] nwinfo = self.network_api.allocate_for_instance( [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] created_port_ids = self._update_ports_for_instance( [ 661.843473] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] with excutils.save_and_reraise_exception(): [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self.force_reraise() [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] raise self.value [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] updated_port = self._update_port( [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] _ensure_no_port_binding_failure(port) [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] raise exception.PortBindingFailed(port_id=port['id']) [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] nova.exception.PortBindingFailed: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. [ 661.843775] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] [ 661.844102] env[63197]: INFO nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Terminating instance [ 661.846655] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Acquiring lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.039399] env[63197]: DEBUG nova.compute.utils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.040842] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.041028] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 662.144063] env[63197]: DEBUG nova.policy [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '94419f4b0e6346ea830b0072ab80f798', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '969bf888642249289ef4a81d5492d698', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 662.338381] env[63197]: DEBUG nova.network.neutron [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.511355] env[63197]: DEBUG nova.network.neutron [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.547637] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.666063] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Successfully created port: 3e179f8d-7685-4f62-a015-abf0b0bea999 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.714437] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Acquiring lock "a90e6999-9aab-4b83-98e4-2b71681d1bb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.714437] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Lock "a90e6999-9aab-4b83-98e4-2b71681d1bb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.018129] env[63197]: DEBUG oslo_concurrency.lockutils [req-581eb6f4-f317-4784-bb9d-c96a6fd700d6 req-9e0cbe95-e1c0-4042-82ee-0a867959b7e6 service nova] Releasing lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.018129] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Acquired lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.018129] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.024960] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d909027-ea2d-4c4b-a6cf-3b652360e92e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.033449] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6e29a5-0875-43ae-850e-884258faf035 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.072875] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85309247-24d2-4c0a-abe4-e5f5426a8038 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.082028] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cccdc0e-b3ec-40d2-be47-fd207ec414f5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.096125] env[63197]: DEBUG nova.compute.provider_tree [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.537845] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.576489] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.600610] env[63197]: DEBUG nova.scheduler.client.report [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.607790] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.608199] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.609180] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.609180] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.609180] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.609180] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.609180] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.609632] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.609632] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.609632] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.609729] env[63197]: DEBUG nova.virt.hardware [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.612827] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a047873a-433e-4a5a-8cc2-ad808f9269a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.622294] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe6eb2f-1e22-41f4-9821-85379dd7075c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.696297] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.829602] env[63197]: DEBUG nova.compute.manager [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Received event network-vif-deleted-3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.829809] env[63197]: DEBUG nova.compute.manager [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Received event network-changed-3e179f8d-7685-4f62-a015-abf0b0bea999 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.830273] env[63197]: DEBUG nova.compute.manager [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Refreshing instance network info cache due to event network-changed-3e179f8d-7685-4f62-a015-abf0b0bea999. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.830273] env[63197]: DEBUG oslo_concurrency.lockutils [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] Acquiring lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.830392] env[63197]: DEBUG oslo_concurrency.lockutils [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] Acquired lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.830543] env[63197]: DEBUG nova.network.neutron [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Refreshing network info cache for port 3e179f8d-7685-4f62-a015-abf0b0bea999 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 664.110477] env[63197]: ERROR nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. [ 664.110477] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.110477] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.110477] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.110477] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.110477] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.110477] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.110477] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.110477] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.110477] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 664.110477] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.110477] env[63197]: ERROR nova.compute.manager raise self.value [ 664.110477] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.110477] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.110477] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.110477] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.111143] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.111143] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.111143] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. [ 664.111143] env[63197]: ERROR nova.compute.manager [ 664.111143] env[63197]: Traceback (most recent call last): [ 664.111143] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.111143] env[63197]: listener.cb(fileno) [ 664.111143] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.111143] env[63197]: result = function(*args, **kwargs) [ 664.111143] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.111143] env[63197]: return func(*args, **kwargs) [ 664.111143] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.111143] env[63197]: raise e [ 664.111143] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.111143] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 664.111143] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.111143] env[63197]: created_port_ids = self._update_ports_for_instance( [ 664.111143] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.111143] env[63197]: with excutils.save_and_reraise_exception(): [ 664.111143] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.111143] env[63197]: self.force_reraise() [ 664.111143] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.111143] env[63197]: raise self.value [ 664.111143] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.111143] env[63197]: updated_port = self._update_port( [ 664.111143] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.111143] env[63197]: _ensure_no_port_binding_failure(port) [ 664.111143] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.111143] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.111909] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. [ 664.111909] env[63197]: Removing descriptor: 18 [ 664.111909] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.112480] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.119249] env[63197]: ERROR nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Traceback (most recent call last): [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] yield resources [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self.driver.spawn(context, instance, image_meta, [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] vm_ref = self.build_virtual_machine(instance, [ 664.119249] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] for vif in network_info: [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return self._sync_wrapper(fn, *args, **kwargs) [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self.wait() [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self[:] = self._gt.wait() [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return self._exit_event.wait() [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.119628] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] result = hub.switch() [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return self.greenlet.switch() [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] result = function(*args, **kwargs) [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return func(*args, **kwargs) [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] raise e [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] nwinfo = self.network_api.allocate_for_instance( [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] created_port_ids = self._update_ports_for_instance( [ 664.120068] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] with excutils.save_and_reraise_exception(): [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self.force_reraise() [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] raise self.value [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] updated_port = self._update_port( [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] _ensure_no_port_binding_failure(port) [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] raise exception.PortBindingFailed(port_id=port['id']) [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] nova.exception.PortBindingFailed: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. [ 664.120422] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] [ 664.120781] env[63197]: INFO nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Terminating instance [ 664.123017] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.370s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.127013] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.198979] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Releasing lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.199423] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 664.199611] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 664.199884] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21e090be-4107-47c8-9fd9-6a2b5a322610 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.823463] env[63197]: DEBUG nova.compute.utils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 664.827626] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 664.827869] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 664.837416] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9e82fa-c7f5-43c9-9966-93b55a068c7c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.850639] env[63197]: DEBUG nova.network.neutron [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.867079] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f2d4315-d533-431a-970a-ea655febf0ab could not be found. [ 664.867193] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.867376] env[63197]: INFO nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Took 0.67 seconds to destroy the instance on the hypervisor. [ 664.867608] env[63197]: DEBUG oslo.service.loopingcall [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.869563] env[63197]: DEBUG nova.compute.manager [-] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.869783] env[63197]: DEBUG nova.network.neutron [-] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.903999] env[63197]: DEBUG nova.network.neutron [-] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.936092] env[63197]: DEBUG nova.policy [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '337bb4e670064920bcaf5014d2b0d83a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a57223cff39b4e8c9a6800c7e53a2cb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 664.946823] env[63197]: DEBUG nova.network.neutron [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.219733] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "bd771a95-98b1-4fdb-b213-42d693e027f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.220047] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "bd771a95-98b1-4fdb-b213-42d693e027f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.238223] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f132042b-c415-4672-a61a-0e255a5955f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.248342] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afe147a-a51e-4492-800c-124df8fd6f71 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.280902] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be02bbc-3cea-4ae0-b60a-0fc5049d4687 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.291300] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dcf2bac-3ef9-409d-85b6-32853d55b803 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.302525] env[63197]: DEBUG nova.compute.provider_tree [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.329034] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.406991] env[63197]: DEBUG nova.network.neutron [-] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.449572] env[63197]: DEBUG oslo_concurrency.lockutils [req-2a5c9cf3-2e89-4b34-a8b2-9c27d79a1eb6 req-f5248f3a-8d02-4e62-8259-54542bb44705 service nova] Releasing lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.450386] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquired lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.450740] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 665.545349] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Successfully created port: ff1cf96a-19b0-402c-9174-5203beb952cf {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.781467] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "d7a2de2b-d14b-437c-93ff-17ee24bd97ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.781737] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "d7a2de2b-d14b-437c-93ff-17ee24bd97ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.805964] env[63197]: DEBUG nova.scheduler.client.report [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.910488] env[63197]: INFO nova.compute.manager [-] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Took 1.04 seconds to deallocate network for instance. [ 665.913555] env[63197]: DEBUG nova.compute.claims [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 665.913953] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.976365] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.102500] env[63197]: DEBUG nova.compute.manager [req-e4f2c975-183f-41df-b1a4-6e2791edba4a req-8456b75e-1b3d-4fdb-a73f-129654fcb87b service nova] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Received event network-vif-deleted-3e179f8d-7685-4f62-a015-abf0b0bea999 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.103948] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.310881] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.189s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.311565] env[63197]: ERROR nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Traceback (most recent call last): [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self.driver.spawn(context, instance, image_meta, [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] vm_ref = self.build_virtual_machine(instance, [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.311565] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] for vif in network_info: [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return self._sync_wrapper(fn, *args, **kwargs) [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self.wait() [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self[:] = self._gt.wait() [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return self._exit_event.wait() [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] result = hub.switch() [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.311889] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return self.greenlet.switch() [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] result = function(*args, **kwargs) [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] return func(*args, **kwargs) [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] raise e [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] nwinfo = self.network_api.allocate_for_instance( [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] created_port_ids = self._update_ports_for_instance( [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] with excutils.save_and_reraise_exception(): [ 666.312220] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] self.force_reraise() [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] raise self.value [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] updated_port = self._update_port( [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] _ensure_no_port_binding_failure(port) [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] raise exception.PortBindingFailed(port_id=port['id']) [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] nova.exception.PortBindingFailed: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. [ 666.312539] env[63197]: ERROR nova.compute.manager [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] [ 666.312842] env[63197]: DEBUG nova.compute.utils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.313690] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.215s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.316162] env[63197]: INFO nova.compute.claims [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.319758] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Build of instance f01a025d-2cb9-4cb9-878a-b05e318c21b9 was re-scheduled: Binding failed for port 44d43c99-1a55-4847-809d-e6abdc64cc3f, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.321130] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.321130] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Acquiring lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.321130] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Acquired lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.321130] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.339120] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.370377] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.370645] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.370805] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.370989] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.371160] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.371640] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.371640] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.371640] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.371794] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.371954] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.372334] env[63197]: DEBUG nova.virt.hardware [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.373232] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478cf857-429c-41eb-b549-95e58dda8a36 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.385067] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e95322c-0aed-4d47-980b-c9c98e64a201 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.558563] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "82919ffb-2b50-4336-9517-c741d259f19e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.558790] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "82919ffb-2b50-4336-9517-c741d259f19e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.608490] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Releasing lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.608490] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 666.608490] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 666.608490] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9aeca80b-2862-439b-8734-945a8a96b6a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.616737] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1975b197-7d84-4e31-8c02-6896b29896f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.640975] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 47da4c8f-a4cb-4d7f-81c0-7c3019169b56 could not be found. [ 666.641313] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.642181] env[63197]: INFO nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Took 0.03 seconds to destroy the instance on the hypervisor. [ 666.642181] env[63197]: DEBUG oslo.service.loopingcall [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.642181] env[63197]: DEBUG nova.compute.manager [-] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.642181] env[63197]: DEBUG nova.network.neutron [-] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.661898] env[63197]: DEBUG nova.network.neutron [-] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.673026] env[63197]: DEBUG nova.compute.manager [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Received event network-changed-ff1cf96a-19b0-402c-9174-5203beb952cf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 666.673026] env[63197]: DEBUG nova.compute.manager [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Refreshing instance network info cache due to event network-changed-ff1cf96a-19b0-402c-9174-5203beb952cf. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 666.673026] env[63197]: DEBUG oslo_concurrency.lockutils [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] Acquiring lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.673026] env[63197]: DEBUG oslo_concurrency.lockutils [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] Acquired lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.673026] env[63197]: DEBUG nova.network.neutron [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Refreshing network info cache for port ff1cf96a-19b0-402c-9174-5203beb952cf {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 666.842463] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.873276] env[63197]: ERROR nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. [ 666.873276] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 666.873276] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.873276] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 666.873276] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.873276] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 666.873276] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.873276] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 666.873276] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.873276] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 666.873276] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.873276] env[63197]: ERROR nova.compute.manager raise self.value [ 666.873276] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.873276] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 666.873276] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.873276] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 666.873762] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.873762] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 666.873762] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. [ 666.873762] env[63197]: ERROR nova.compute.manager [ 666.873762] env[63197]: Traceback (most recent call last): [ 666.873762] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 666.873762] env[63197]: listener.cb(fileno) [ 666.873762] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.873762] env[63197]: result = function(*args, **kwargs) [ 666.873762] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.873762] env[63197]: return func(*args, **kwargs) [ 666.873762] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.873762] env[63197]: raise e [ 666.873762] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.873762] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 666.873762] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.873762] env[63197]: created_port_ids = self._update_ports_for_instance( [ 666.873762] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.873762] env[63197]: with excutils.save_and_reraise_exception(): [ 666.873762] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.873762] env[63197]: self.force_reraise() [ 666.873762] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.873762] env[63197]: raise self.value [ 666.873762] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.873762] env[63197]: updated_port = self._update_port( [ 666.873762] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.873762] env[63197]: _ensure_no_port_binding_failure(port) [ 666.873762] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.873762] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 666.874560] env[63197]: nova.exception.PortBindingFailed: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. [ 666.874560] env[63197]: Removing descriptor: 19 [ 666.874560] env[63197]: ERROR nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Traceback (most recent call last): [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] yield resources [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self.driver.spawn(context, instance, image_meta, [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.874560] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] vm_ref = self.build_virtual_machine(instance, [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] for vif in network_info: [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return self._sync_wrapper(fn, *args, **kwargs) [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self.wait() [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self[:] = self._gt.wait() [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return self._exit_event.wait() [ 666.874891] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] result = hub.switch() [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return self.greenlet.switch() [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] result = function(*args, **kwargs) [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return func(*args, **kwargs) [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] raise e [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] nwinfo = self.network_api.allocate_for_instance( [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.875231] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] created_port_ids = self._update_ports_for_instance( [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] with excutils.save_and_reraise_exception(): [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self.force_reraise() [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] raise self.value [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] updated_port = self._update_port( [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] _ensure_no_port_binding_failure(port) [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.875562] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] raise exception.PortBindingFailed(port_id=port['id']) [ 666.875850] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] nova.exception.PortBindingFailed: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. [ 666.875850] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] [ 666.875850] env[63197]: INFO nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Terminating instance [ 666.876575] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Acquiring lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.913993] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.165302] env[63197]: DEBUG nova.network.neutron [-] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.193250] env[63197]: DEBUG nova.network.neutron [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.283214] env[63197]: DEBUG nova.network.neutron [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.418000] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Releasing lock "refresh_cache-f01a025d-2cb9-4cb9-878a-b05e318c21b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.418000] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.418000] env[63197]: DEBUG nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.418000] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.432674] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.667771] env[63197]: INFO nova.compute.manager [-] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Took 1.03 seconds to deallocate network for instance. [ 667.670863] env[63197]: DEBUG nova.compute.claims [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.671093] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.742025] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec1fdf0-a1b8-475a-a28d-d77cb24355f9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.749749] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5561be69-ca23-4be4-a387-6b492a1d22f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.789386] env[63197]: DEBUG oslo_concurrency.lockutils [req-59351ef6-b539-4ba4-bed0-863b3c7fd9f7 req-c9af24c2-012c-450d-b9cc-fc75088dd128 service nova] Releasing lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.790094] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Acquired lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.790254] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.791889] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f009009e-0793-4190-b808-eb0a03425002 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.801593] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251d0820-372d-4221-ad32-3bf8f6a93465 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.814769] env[63197]: DEBUG nova.compute.provider_tree [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.935614] env[63197]: DEBUG nova.network.neutron [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.317345] env[63197]: DEBUG nova.scheduler.client.report [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.321090] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.430847] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.439308] env[63197]: INFO nova.compute.manager [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] [instance: f01a025d-2cb9-4cb9-878a-b05e318c21b9] Took 1.02 seconds to deallocate network for instance. [ 668.794420] env[63197]: DEBUG nova.compute.manager [req-195c6d5d-2986-4634-8300-af5f49e19236 req-6ceb9800-aad7-4c73-8d69-ac06037c34a1 service nova] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Received event network-vif-deleted-ff1cf96a-19b0-402c-9174-5203beb952cf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 668.824243] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.824713] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.831025] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.217s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.831025] env[63197]: DEBUG nova.objects.instance [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63197) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 668.932556] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Releasing lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.932991] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 668.933190] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 668.933473] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5261132-8ab9-4428-9ab0-29c824002bba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.948361] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7893907-c399-4445-9b33-884b93402955 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.969492] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6abf52eb-a394-477f-95e6-87cddecd4bd0 could not be found. [ 668.969743] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 668.969924] env[63197]: INFO nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 668.970221] env[63197]: DEBUG oslo.service.loopingcall [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.970449] env[63197]: DEBUG nova.compute.manager [-] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.970578] env[63197]: DEBUG nova.network.neutron [-] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.993716] env[63197]: DEBUG nova.network.neutron [-] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.333973] env[63197]: DEBUG nova.compute.utils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.336767] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 669.336925] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 669.391736] env[63197]: DEBUG nova.policy [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08c1762c485d40288a8aa2532901c44b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97d57fe11046425cbb88884a986dd009', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 669.497551] env[63197]: INFO nova.scheduler.client.report [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Deleted allocations for instance f01a025d-2cb9-4cb9-878a-b05e318c21b9 [ 669.507799] env[63197]: DEBUG nova.network.neutron [-] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.787776] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Successfully created port: 6ef74f8b-974e-4ab6-b725-a3c93b507d19 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.840723] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.848264] env[63197]: DEBUG oslo_concurrency.lockutils [None req-38608412-1f9b-461c-ae41-ba39039fcdab tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.849529] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.783s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.015565] env[63197]: DEBUG oslo_concurrency.lockutils [None req-622f9c1b-e1eb-4ca2-9b4c-15f88aa51c90 tempest-ServersTestManualDisk-1257409225 tempest-ServersTestManualDisk-1257409225-project-member] Lock "f01a025d-2cb9-4cb9-878a-b05e318c21b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.159s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.016097] env[63197]: INFO nova.compute.manager [-] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Took 1.05 seconds to deallocate network for instance. [ 670.020083] env[63197]: DEBUG nova.compute.claims [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.021029] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.520111] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.793489] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 670.793489] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 670.796512] env[63197]: DEBUG nova.compute.manager [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Received event network-changed-6ef74f8b-974e-4ab6-b725-a3c93b507d19 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.796875] env[63197]: DEBUG nova.compute.manager [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Refreshing instance network info cache due to event network-changed-6ef74f8b-974e-4ab6-b725-a3c93b507d19. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 670.797731] env[63197]: DEBUG oslo_concurrency.lockutils [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] Acquiring lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.798061] env[63197]: DEBUG oslo_concurrency.lockutils [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] Acquired lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.798461] env[63197]: DEBUG nova.network.neutron [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Refreshing network info cache for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 670.845295] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb282d58-cf79-4d53-9866-fc209011f87a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.854967] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a73f157-cfb0-4185-b3a5-34039ae82a4e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.890816] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 670.896190] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666cb8ff-5ca6-4ee1-8852-d38e97f6314c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.904829] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a57faea-784b-403c-a7a0-507625948372 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.920734] env[63197]: DEBUG nova.compute.provider_tree [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.934615] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 670.935670] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 670.935670] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 670.935670] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 670.935670] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 670.935670] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 670.936417] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 670.936417] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 670.936417] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 670.936417] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 670.936714] env[63197]: DEBUG nova.virt.hardware [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.937328] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6261a37d-3af3-4d4c-ae3d-2390e1fbe544 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.950196] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601e6f35-e827-4a95-9cbe-91217c713647 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.044225] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.267775] env[63197]: ERROR nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. [ 671.267775] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 671.267775] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.267775] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 671.267775] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.267775] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 671.267775] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.267775] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 671.267775] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.267775] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 671.267775] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.267775] env[63197]: ERROR nova.compute.manager raise self.value [ 671.267775] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.267775] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 671.267775] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.267775] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 671.268252] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.268252] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 671.268252] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. [ 671.268252] env[63197]: ERROR nova.compute.manager [ 671.268252] env[63197]: Traceback (most recent call last): [ 671.268252] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 671.268252] env[63197]: listener.cb(fileno) [ 671.268252] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.268252] env[63197]: result = function(*args, **kwargs) [ 671.268252] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.268252] env[63197]: return func(*args, **kwargs) [ 671.268252] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.268252] env[63197]: raise e [ 671.268252] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.268252] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 671.268252] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.268252] env[63197]: created_port_ids = self._update_ports_for_instance( [ 671.268252] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.268252] env[63197]: with excutils.save_and_reraise_exception(): [ 671.268252] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.268252] env[63197]: self.force_reraise() [ 671.268252] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.268252] env[63197]: raise self.value [ 671.268252] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.268252] env[63197]: updated_port = self._update_port( [ 671.268252] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.268252] env[63197]: _ensure_no_port_binding_failure(port) [ 671.268252] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.268252] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.268919] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. [ 671.268919] env[63197]: Removing descriptor: 18 [ 671.268919] env[63197]: ERROR nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Traceback (most recent call last): [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] yield resources [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self.driver.spawn(context, instance, image_meta, [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.268919] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] vm_ref = self.build_virtual_machine(instance, [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] for vif in network_info: [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return self._sync_wrapper(fn, *args, **kwargs) [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self.wait() [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self[:] = self._gt.wait() [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return self._exit_event.wait() [ 671.269255] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] result = hub.switch() [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return self.greenlet.switch() [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] result = function(*args, **kwargs) [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return func(*args, **kwargs) [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] raise e [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] nwinfo = self.network_api.allocate_for_instance( [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.269577] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] created_port_ids = self._update_ports_for_instance( [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] with excutils.save_and_reraise_exception(): [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self.force_reraise() [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] raise self.value [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] updated_port = self._update_port( [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] _ensure_no_port_binding_failure(port) [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.269904] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] raise exception.PortBindingFailed(port_id=port['id']) [ 671.270249] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] nova.exception.PortBindingFailed: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. [ 671.270249] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] [ 671.270249] env[63197]: INFO nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Terminating instance [ 671.273365] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Acquiring lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.306146] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 671.306308] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 671.306425] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Rebuilding the list of instances to heal {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 671.328191] env[63197]: DEBUG nova.network.neutron [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.428628] env[63197]: DEBUG nova.scheduler.client.report [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.467102] env[63197]: DEBUG nova.network.neutron [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.810570] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 671.810850] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 671.810924] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 671.811059] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 671.811313] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 671.841958] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "refresh_cache-a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.842120] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquired lock "refresh_cache-a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.842265] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Forcefully refreshing network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 671.842418] env[63197]: DEBUG nova.objects.instance [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lazy-loading 'info_cache' on Instance uuid a066e6e1-12a4-455b-87cc-e3d6a6a17c1f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 671.933904] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.084s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.934590] env[63197]: ERROR nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Traceback (most recent call last): [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self.driver.spawn(context, instance, image_meta, [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] vm_ref = self.build_virtual_machine(instance, [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.934590] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] for vif in network_info: [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return self._sync_wrapper(fn, *args, **kwargs) [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self.wait() [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self[:] = self._gt.wait() [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return self._exit_event.wait() [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] result = hub.switch() [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.934944] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return self.greenlet.switch() [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] result = function(*args, **kwargs) [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] return func(*args, **kwargs) [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] raise e [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] nwinfo = self.network_api.allocate_for_instance( [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] created_port_ids = self._update_ports_for_instance( [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] with excutils.save_and_reraise_exception(): [ 671.935381] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] self.force_reraise() [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] raise self.value [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] updated_port = self._update_port( [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] _ensure_no_port_binding_failure(port) [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] raise exception.PortBindingFailed(port_id=port['id']) [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] nova.exception.PortBindingFailed: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. [ 671.936014] env[63197]: ERROR nova.compute.manager [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] [ 671.936330] env[63197]: DEBUG nova.compute.utils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.936622] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.982s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.941097] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Build of instance b52f3d4b-7699-484d-8b4e-84393848fb37 was re-scheduled: Binding failed for port 9223f478-9a65-4de3-ad95-dcd4a96500c4, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 671.941806] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.942086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Acquiring lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.942298] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Acquired lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.942491] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.972728] env[63197]: DEBUG oslo_concurrency.lockutils [req-a2e2c789-1d39-4a89-a19b-508f60661595 req-2d14c670-f98b-4450-b75e-7c3348349d1a service nova] Releasing lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.973148] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Acquired lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.973325] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.462047] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.506343] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.602528] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.604387] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.829446] env[63197]: DEBUG nova.compute.manager [req-0d864583-c7c9-4d95-8085-7dad39ff86c8 req-3acff82d-e9f7-48a3-b619-b22dc9d99e65 service nova] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Received event network-vif-deleted-6ef74f8b-974e-4ab6-b725-a3c93b507d19 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.864423] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.886444] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ab4377-dc03-4a17-86dc-36980670ecd2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.894641] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6dcd10c-946e-4bb9-8083-195ca039c0f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.926105] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ce11c6-23f6-4987-afd0-89ea7fef8997 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.933448] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3be103b-413e-4d3d-a51f-d7ae3db77e8e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.946550] env[63197]: DEBUG nova.compute.provider_tree [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.106518] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Releasing lock "refresh_cache-b52f3d4b-7699-484d-8b4e-84393848fb37" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.106774] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 673.106954] env[63197]: DEBUG nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.107134] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.110706] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Releasing lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.111249] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 673.111465] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 673.111605] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c3cd130-39ae-4537-84e4-3673d92e540b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.120986] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f229c17c-988c-41af-b6be-665ae47656d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.131826] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.144412] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0d7e1031-896b-4ec3-9da0-74d65b1ef78a could not be found. [ 673.144628] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 673.144815] env[63197]: INFO nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 673.145054] env[63197]: DEBUG oslo.service.loopingcall [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.145269] env[63197]: DEBUG nova.compute.manager [-] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.145399] env[63197]: DEBUG nova.network.neutron [-] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.160594] env[63197]: DEBUG nova.network.neutron [-] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.431103] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.449874] env[63197]: DEBUG nova.scheduler.client.report [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.634547] env[63197]: DEBUG nova.network.neutron [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.662734] env[63197]: DEBUG nova.network.neutron [-] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.933904] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Releasing lock "refresh_cache-a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.934193] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Updated the network info_cache for instance {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 673.934269] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.934442] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.934657] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.934876] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.934957] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.935119] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.935250] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 673.935425] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.954745] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.018s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.955378] env[63197]: ERROR nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Traceback (most recent call last): [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self.driver.spawn(context, instance, image_meta, [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] vm_ref = self.build_virtual_machine(instance, [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.955378] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] for vif in network_info: [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return self._sync_wrapper(fn, *args, **kwargs) [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self.wait() [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self[:] = self._gt.wait() [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return self._exit_event.wait() [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] result = hub.switch() [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 673.955797] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return self.greenlet.switch() [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] result = function(*args, **kwargs) [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] return func(*args, **kwargs) [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] raise e [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] nwinfo = self.network_api.allocate_for_instance( [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] created_port_ids = self._update_ports_for_instance( [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] with excutils.save_and_reraise_exception(): [ 673.956355] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] self.force_reraise() [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] raise self.value [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] updated_port = self._update_port( [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] _ensure_no_port_binding_failure(port) [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] raise exception.PortBindingFailed(port_id=port['id']) [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] nova.exception.PortBindingFailed: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. [ 673.957223] env[63197]: ERROR nova.compute.manager [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] [ 673.957665] env[63197]: DEBUG nova.compute.utils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.957665] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.987s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.957665] env[63197]: DEBUG nova.objects.instance [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lazy-loading 'resources' on Instance uuid a066e6e1-12a4-455b-87cc-e3d6a6a17c1f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 673.960797] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Build of instance a18802fb-2a73-4bb5-aab9-5a7eb599a5a6 was re-scheduled: Binding failed for port d4c6030d-61e8-454b-8aea-337e0cf65ba2, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.961375] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.961694] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Acquiring lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.961897] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Acquired lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.962115] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.137655] env[63197]: INFO nova.compute.manager [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] [instance: b52f3d4b-7699-484d-8b4e-84393848fb37] Took 1.03 seconds to deallocate network for instance. [ 674.168347] env[63197]: INFO nova.compute.manager [-] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Took 1.02 seconds to deallocate network for instance. [ 674.171543] env[63197]: DEBUG nova.compute.claims [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 674.172575] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.438826] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.494022] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.593743] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.951854] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b135ab8-5ad2-41cc-b787-5361773be395 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.961928] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af7130a-e401-477a-8f9d-ce011c551ee8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.993968] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c58c76-d3f5-404d-a9a6-7165a80a9f51 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.001727] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1c0956-cdc7-4676-8841-afc0f468dc19 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.016457] env[63197]: DEBUG nova.compute.provider_tree [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.098192] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Releasing lock "refresh_cache-a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.098387] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 675.098573] env[63197]: DEBUG nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.098744] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.121483] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.176168] env[63197]: INFO nova.scheduler.client.report [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Deleted allocations for instance b52f3d4b-7699-484d-8b4e-84393848fb37 [ 675.522674] env[63197]: DEBUG nova.scheduler.client.report [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.624646] env[63197]: DEBUG nova.network.neutron [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.685947] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0d69b129-8e7a-4e54-a232-2013c96b0f60 tempest-ServersTestJSON-401115467 tempest-ServersTestJSON-401115467-project-member] Lock "b52f3d4b-7699-484d-8b4e-84393848fb37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.459s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.029962] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.071s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.031887] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.850s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.034312] env[63197]: INFO nova.compute.claims [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.060433] env[63197]: INFO nova.scheduler.client.report [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Deleted allocations for instance a066e6e1-12a4-455b-87cc-e3d6a6a17c1f [ 676.128517] env[63197]: INFO nova.compute.manager [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] [instance: a18802fb-2a73-4bb5-aab9-5a7eb599a5a6] Took 1.03 seconds to deallocate network for instance. [ 676.188851] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 676.433843] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Acquiring lock "66af0137-22d4-4670-8225-8d9b9dd2884a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.434396] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Lock "66af0137-22d4-4670-8225-8d9b9dd2884a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.573078] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a2187f07-4167-4b70-b897-d35598817658 tempest-ServerShowV257Test-470171500 tempest-ServerShowV257Test-470171500-project-member] Lock "a066e6e1-12a4-455b-87cc-e3d6a6a17c1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.328s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.717968] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.171438] env[63197]: INFO nova.scheduler.client.report [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Deleted allocations for instance a18802fb-2a73-4bb5-aab9-5a7eb599a5a6 [ 677.470567] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5f2689-d6b9-43fb-87ea-830ad2cc6b17 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.479033] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41eb4f51-255d-4cf6-946b-41a54a3611c3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.512961] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a88c45-da14-4a87-a946-e5ec84782078 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.520993] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8348e9d0-b43c-4f00-ae95-f5883985fc0e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.535186] env[63197]: DEBUG nova.compute.provider_tree [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.685019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4587a1f1-2d92-4c0c-b9cf-130a9c9e366a tempest-ServersTestFqdnHostnames-715351491 tempest-ServersTestFqdnHostnames-715351491-project-member] Lock "a18802fb-2a73-4bb5-aab9-5a7eb599a5a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.556s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.039457] env[63197]: DEBUG nova.scheduler.client.report [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.251973] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 678.543912] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.544470] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.547337] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.157s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.548843] env[63197]: INFO nova.compute.claims [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.735431] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.057887] env[63197]: DEBUG nova.compute.utils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.059429] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 679.059603] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 679.220495] env[63197]: DEBUG nova.policy [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c91dabb95115428cac178b75dbd4a8ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e2f93ce59984a2f8ced7987b99df5cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.568606] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.042030] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Successfully created port: a966cac4-41f3-4e36-ba90-8c0a1c2b5765 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.192687] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c9857eb-6e69-4335-8699-386f58fbf40e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.200302] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d16d15-1523-41f7-99c7-0f42f9f197c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.233193] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2248b326-dcee-40ae-aa5e-d6e39c621295 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.241629] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada5ca2e-5cf7-4699-9850-55c3b7069c21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.255444] env[63197]: DEBUG nova.compute.provider_tree [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.590034] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.625236] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.625490] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.625692] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.625918] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.626186] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.626387] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.626634] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.626827] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.627038] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.627238] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.627448] env[63197]: DEBUG nova.virt.hardware [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.628405] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df1ce29-80ef-40ab-9e65-b1eeb32b9ed9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.636279] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b6010e-8174-4c70-841a-fd39ce921376 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.759893] env[63197]: DEBUG nova.scheduler.client.report [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.266358] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.719s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.266923] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 681.271877] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.012s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.275028] env[63197]: INFO nova.compute.claims [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.451930] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Acquiring lock "0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.452194] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Lock "0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.490067] env[63197]: DEBUG nova.compute.manager [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Received event network-changed-a966cac4-41f3-4e36-ba90-8c0a1c2b5765 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 681.490373] env[63197]: DEBUG nova.compute.manager [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Refreshing instance network info cache due to event network-changed-a966cac4-41f3-4e36-ba90-8c0a1c2b5765. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 681.490698] env[63197]: DEBUG oslo_concurrency.lockutils [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] Acquiring lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.491033] env[63197]: DEBUG oslo_concurrency.lockutils [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] Acquired lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.491510] env[63197]: DEBUG nova.network.neutron [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Refreshing network info cache for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 681.693629] env[63197]: ERROR nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. [ 681.693629] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 681.693629] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.693629] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 681.693629] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.693629] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 681.693629] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.693629] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 681.693629] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.693629] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 681.693629] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.693629] env[63197]: ERROR nova.compute.manager raise self.value [ 681.693629] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.693629] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 681.693629] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.693629] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 681.694309] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.694309] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 681.694309] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. [ 681.694309] env[63197]: ERROR nova.compute.manager [ 681.694309] env[63197]: Traceback (most recent call last): [ 681.694309] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 681.694309] env[63197]: listener.cb(fileno) [ 681.694309] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.694309] env[63197]: result = function(*args, **kwargs) [ 681.694309] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.694309] env[63197]: return func(*args, **kwargs) [ 681.694309] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.694309] env[63197]: raise e [ 681.694309] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.694309] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 681.694309] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.694309] env[63197]: created_port_ids = self._update_ports_for_instance( [ 681.694309] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.694309] env[63197]: with excutils.save_and_reraise_exception(): [ 681.694309] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.694309] env[63197]: self.force_reraise() [ 681.694309] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.694309] env[63197]: raise self.value [ 681.694309] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.694309] env[63197]: updated_port = self._update_port( [ 681.694309] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.694309] env[63197]: _ensure_no_port_binding_failure(port) [ 681.694309] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.694309] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 681.695103] env[63197]: nova.exception.PortBindingFailed: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. [ 681.695103] env[63197]: Removing descriptor: 19 [ 681.695103] env[63197]: ERROR nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Traceback (most recent call last): [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] yield resources [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self.driver.spawn(context, instance, image_meta, [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.695103] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] vm_ref = self.build_virtual_machine(instance, [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] for vif in network_info: [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return self._sync_wrapper(fn, *args, **kwargs) [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self.wait() [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self[:] = self._gt.wait() [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return self._exit_event.wait() [ 681.695447] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] result = hub.switch() [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return self.greenlet.switch() [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] result = function(*args, **kwargs) [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return func(*args, **kwargs) [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] raise e [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] nwinfo = self.network_api.allocate_for_instance( [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.695936] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] created_port_ids = self._update_ports_for_instance( [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] with excutils.save_and_reraise_exception(): [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self.force_reraise() [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] raise self.value [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] updated_port = self._update_port( [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] _ensure_no_port_binding_failure(port) [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.696658] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] raise exception.PortBindingFailed(port_id=port['id']) [ 681.697964] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] nova.exception.PortBindingFailed: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. [ 681.697964] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] [ 681.697964] env[63197]: INFO nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Terminating instance [ 681.700665] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.784289] env[63197]: DEBUG nova.compute.utils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 681.785726] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 681.785726] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 681.833010] env[63197]: DEBUG nova.policy [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '48823701def34763a644035831812150', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f369e4b355047a4be627fa8500e708d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.013251] env[63197]: DEBUG nova.network.neutron [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.182996] env[63197]: DEBUG nova.network.neutron [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.256876] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Successfully created port: aafa3f61-f429-429a-ad97-fcb08a5df883 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 682.291919] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 682.685954] env[63197]: DEBUG oslo_concurrency.lockutils [req-6df11df9-5434-499e-a532-744793d911f6 req-9c1ef0e6-124b-4e4b-9869-e5a1e25c9742 service nova] Releasing lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.686418] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquired lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.686603] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.825149] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eef0aa6-cb64-497b-8a5c-d4e5ba0b99ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.835807] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5556fa-930c-41d3-848e-59b92d507645 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.872459] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82971ed4-dfad-4619-8b5e-a5abbe7d94b2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.880862] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6c6987-a433-448d-a276-1576cd1743cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.894619] env[63197]: DEBUG nova.compute.provider_tree [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.213198] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.306743] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 683.314187] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.340985] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 683.341268] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 683.341428] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.341607] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 683.341752] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.342176] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 683.342176] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 683.342267] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 683.342536] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 683.342598] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 683.342726] env[63197]: DEBUG nova.virt.hardware [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 683.345239] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b99d21-b908-495a-994a-4ca2e55aa9fd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.353271] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741fce23-3a7a-434f-892c-9ee8ba75dbe3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.398211] env[63197]: DEBUG nova.scheduler.client.report [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.797051] env[63197]: DEBUG nova.compute.manager [req-efabf00e-7c20-40d8-ba13-3358d7ee12d9 req-ec3a80dd-0740-4597-8a81-b6ae1a4b4dab service nova] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Received event network-vif-deleted-a966cac4-41f3-4e36-ba90-8c0a1c2b5765 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 683.822307] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Releasing lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.822764] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.823032] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.823249] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f24dad43-122d-40e4-8f9b-fdb93d382736 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.832800] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87bdc31-34e8-4ef4-9640-92d45307cc61 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.856875] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 11c6fcd4-87d3-4ceb-98eb-a46ca2910584 could not be found. [ 683.856875] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.856875] env[63197]: INFO nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Took 0.03 seconds to destroy the instance on the hypervisor. [ 683.856875] env[63197]: DEBUG oslo.service.loopingcall [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.856875] env[63197]: DEBUG nova.compute.manager [-] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.856875] env[63197]: DEBUG nova.network.neutron [-] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.882572] env[63197]: DEBUG nova.network.neutron [-] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.908738] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.908738] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 683.909823] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.996s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.054961] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.055952] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.388186] env[63197]: DEBUG nova.network.neutron [-] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.417115] env[63197]: ERROR nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. [ 684.417115] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.417115] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.417115] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.417115] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.417115] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.417115] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.417115] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.417115] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.417115] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 684.417115] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.417115] env[63197]: ERROR nova.compute.manager raise self.value [ 684.417115] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.417115] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.417115] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.417115] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.417516] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.417516] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.417516] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. [ 684.417516] env[63197]: ERROR nova.compute.manager [ 684.417516] env[63197]: Traceback (most recent call last): [ 684.417516] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.417516] env[63197]: listener.cb(fileno) [ 684.417516] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.417516] env[63197]: result = function(*args, **kwargs) [ 684.417516] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.417516] env[63197]: return func(*args, **kwargs) [ 684.417516] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.417516] env[63197]: raise e [ 684.417516] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.417516] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 684.417516] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.417516] env[63197]: created_port_ids = self._update_ports_for_instance( [ 684.417516] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.417516] env[63197]: with excutils.save_and_reraise_exception(): [ 684.417516] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.417516] env[63197]: self.force_reraise() [ 684.417516] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.417516] env[63197]: raise self.value [ 684.417516] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.417516] env[63197]: updated_port = self._update_port( [ 684.417516] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.417516] env[63197]: _ensure_no_port_binding_failure(port) [ 684.417516] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.417516] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.418188] env[63197]: nova.exception.PortBindingFailed: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. [ 684.418188] env[63197]: Removing descriptor: 19 [ 684.418188] env[63197]: DEBUG nova.compute.utils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.427738] env[63197]: ERROR nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Traceback (most recent call last): [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] yield resources [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self.driver.spawn(context, instance, image_meta, [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] vm_ref = self.build_virtual_machine(instance, [ 684.427738] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] for vif in network_info: [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return self._sync_wrapper(fn, *args, **kwargs) [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self.wait() [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self[:] = self._gt.wait() [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return self._exit_event.wait() [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.428367] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] result = hub.switch() [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return self.greenlet.switch() [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] result = function(*args, **kwargs) [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return func(*args, **kwargs) [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] raise e [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] nwinfo = self.network_api.allocate_for_instance( [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] created_port_ids = self._update_ports_for_instance( [ 684.428897] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] with excutils.save_and_reraise_exception(): [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self.force_reraise() [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] raise self.value [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] updated_port = self._update_port( [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] _ensure_no_port_binding_failure(port) [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] raise exception.PortBindingFailed(port_id=port['id']) [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] nova.exception.PortBindingFailed: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. [ 684.429424] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] [ 684.429889] env[63197]: INFO nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Terminating instance [ 684.429935] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 684.430105] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 684.432012] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Acquiring lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.432302] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Acquired lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.432388] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.493315] env[63197]: DEBUG nova.policy [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f43a1f72c46efa2f36553d1a768ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d97b713fdcb4cd58c6bea06ec23a441', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.890485] env[63197]: INFO nova.compute.manager [-] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Took 1.03 seconds to deallocate network for instance. [ 684.895726] env[63197]: DEBUG nova.compute.claims [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.895726] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.908106] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-becf6fae-10ce-4470-8b77-3952e5d92183 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.914895] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f097f0fa-4ff9-4092-bcfd-b6a781ad1c25 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.944767] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 684.951171] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b84c1f7-9b62-4531-8708-63c5b76ee902 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.959333] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cab4d74-0ff0-4442-86ac-7f89565b4785 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.978228] env[63197]: DEBUG nova.compute.provider_tree [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.981013] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.043622] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Successfully created port: a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.083668] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.487292] env[63197]: DEBUG nova.scheduler.client.report [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.589328] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Releasing lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.590491] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 685.590491] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 685.590491] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94330ecd-0635-41f4-ba52-6805b28866b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.603375] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66f3f35-a437-4905-9aa1-224c497f6781 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.636485] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 68b27a14-f52b-4d89-9cdc-be0529428eb7 could not be found. [ 685.636857] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.636981] env[63197]: INFO nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Took 0.05 seconds to destroy the instance on the hypervisor. [ 685.637302] env[63197]: DEBUG oslo.service.loopingcall [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 685.637602] env[63197]: DEBUG nova.compute.manager [-] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.637727] env[63197]: DEBUG nova.network.neutron [-] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.672100] env[63197]: DEBUG nova.network.neutron [-] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.856043] env[63197]: DEBUG nova.compute.manager [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Received event network-changed-aafa3f61-f429-429a-ad97-fcb08a5df883 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.856243] env[63197]: DEBUG nova.compute.manager [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Refreshing instance network info cache due to event network-changed-aafa3f61-f429-429a-ad97-fcb08a5df883. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 685.857333] env[63197]: DEBUG oslo_concurrency.lockutils [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] Acquiring lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.857333] env[63197]: DEBUG oslo_concurrency.lockutils [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] Acquired lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.857333] env[63197]: DEBUG nova.network.neutron [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Refreshing network info cache for port aafa3f61-f429-429a-ad97-fcb08a5df883 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 685.960816] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 685.983916] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 685.984182] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 685.984507] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.984605] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 685.984671] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.984862] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 685.985022] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 685.985181] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 685.985348] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 685.985506] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 685.985674] env[63197]: DEBUG nova.virt.hardware [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 685.986547] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9528f7-4a52-474f-bbf0-eb0e740c06fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.995149] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.085s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.996111] env[63197]: ERROR nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Traceback (most recent call last): [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self.driver.spawn(context, instance, image_meta, [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] vm_ref = self.build_virtual_machine(instance, [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.996111] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] for vif in network_info: [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return self._sync_wrapper(fn, *args, **kwargs) [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self.wait() [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self[:] = self._gt.wait() [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return self._exit_event.wait() [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] result = hub.switch() [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.996443] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return self.greenlet.switch() [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] result = function(*args, **kwargs) [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] return func(*args, **kwargs) [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] raise e [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] nwinfo = self.network_api.allocate_for_instance( [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] created_port_ids = self._update_ports_for_instance( [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] with excutils.save_and_reraise_exception(): [ 685.996771] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] self.force_reraise() [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] raise self.value [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] updated_port = self._update_port( [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] _ensure_no_port_binding_failure(port) [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] raise exception.PortBindingFailed(port_id=port['id']) [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] nova.exception.PortBindingFailed: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. [ 685.997123] env[63197]: ERROR nova.compute.manager [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] [ 685.997401] env[63197]: DEBUG nova.compute.utils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.997635] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.327s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.001308] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b52f80-6629-478b-a3dd-7f63756df69e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.005441] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Build of instance 4f2d4315-d533-431a-970a-ea655febf0ab was re-scheduled: Binding failed for port 3cdb5fda-1a1a-4ab6-990c-5a69cf93ec80, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 686.005894] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 686.006945] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Acquiring lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.006945] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Acquired lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.006945] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.176388] env[63197]: DEBUG nova.network.neutron [-] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.375152] env[63197]: DEBUG nova.network.neutron [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.466464] env[63197]: ERROR nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. [ 686.466464] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.466464] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.466464] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 686.466464] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.466464] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 686.466464] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.466464] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 686.466464] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.466464] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 686.466464] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.466464] env[63197]: ERROR nova.compute.manager raise self.value [ 686.466464] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.466464] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 686.466464] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.466464] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 686.467096] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.467096] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 686.467096] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. [ 686.467096] env[63197]: ERROR nova.compute.manager [ 686.467096] env[63197]: Traceback (most recent call last): [ 686.467096] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 686.467096] env[63197]: listener.cb(fileno) [ 686.467096] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.467096] env[63197]: result = function(*args, **kwargs) [ 686.467096] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.467096] env[63197]: return func(*args, **kwargs) [ 686.467096] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.467096] env[63197]: raise e [ 686.467096] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.467096] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 686.467096] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.467096] env[63197]: created_port_ids = self._update_ports_for_instance( [ 686.467096] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.467096] env[63197]: with excutils.save_and_reraise_exception(): [ 686.467096] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.467096] env[63197]: self.force_reraise() [ 686.467096] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.467096] env[63197]: raise self.value [ 686.467096] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.467096] env[63197]: updated_port = self._update_port( [ 686.467096] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.467096] env[63197]: _ensure_no_port_binding_failure(port) [ 686.467096] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.467096] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 686.469735] env[63197]: nova.exception.PortBindingFailed: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. [ 686.469735] env[63197]: Removing descriptor: 19 [ 686.469735] env[63197]: ERROR nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Traceback (most recent call last): [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] yield resources [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self.driver.spawn(context, instance, image_meta, [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.469735] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] vm_ref = self.build_virtual_machine(instance, [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] for vif in network_info: [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return self._sync_wrapper(fn, *args, **kwargs) [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self.wait() [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self[:] = self._gt.wait() [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return self._exit_event.wait() [ 686.470270] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] result = hub.switch() [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return self.greenlet.switch() [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] result = function(*args, **kwargs) [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return func(*args, **kwargs) [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] raise e [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] nwinfo = self.network_api.allocate_for_instance( [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.470628] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] created_port_ids = self._update_ports_for_instance( [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] with excutils.save_and_reraise_exception(): [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self.force_reraise() [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] raise self.value [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] updated_port = self._update_port( [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] _ensure_no_port_binding_failure(port) [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.470921] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] raise exception.PortBindingFailed(port_id=port['id']) [ 686.471210] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] nova.exception.PortBindingFailed: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. [ 686.471210] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] [ 686.471210] env[63197]: INFO nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Terminating instance [ 686.473034] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.473034] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.473034] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.491390] env[63197]: DEBUG nova.network.neutron [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.533414] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.641666] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.675819] env[63197]: INFO nova.compute.manager [-] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Took 1.04 seconds to deallocate network for instance. [ 686.678633] env[63197]: DEBUG nova.compute.claims [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 686.678897] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.914077] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ef77e2-fbce-4c84-9705-d5661415d0ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.921794] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e2567c-25c9-42ff-a3b5-44566da0940a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.952256] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b131ed3d-9a7a-48a9-842e-109b488625e2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.960148] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdf2061-426b-40ac-9cfc-f9b97eec7145 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.973395] env[63197]: DEBUG nova.compute.provider_tree [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.993703] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.998786] env[63197]: DEBUG oslo_concurrency.lockutils [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] Releasing lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.999114] env[63197]: DEBUG nova.compute.manager [req-2bf0be61-c56e-4528-b65c-c38b0785aee7 req-d3176460-91ba-4072-9fa2-2ffe72be8d38 service nova] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Received event network-vif-deleted-aafa3f61-f429-429a-ad97-fcb08a5df883 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.004669] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "7786e5f4-a7c9-4c14-b407-d2eda85bf37a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.004889] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "7786e5f4-a7c9-4c14-b407-d2eda85bf37a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.032734] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "5128f0c1-da4a-4eb4-9dde-23f830800512" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.032959] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "5128f0c1-da4a-4eb4-9dde-23f830800512" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.061406] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "06167654-0ac8-4b04-8eb7-071eac0894ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.061629] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "06167654-0ac8-4b04-8eb7-071eac0894ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.090076] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.144684] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Releasing lock "refresh_cache-4f2d4315-d533-431a-970a-ea655febf0ab" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.144925] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.145131] env[63197]: DEBUG nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.145295] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.168076] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.478035] env[63197]: DEBUG nova.scheduler.client.report [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.593640] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.594082] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.594288] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.594585] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f4d6e51-9d3c-4522-9e67-fd767c00b1c8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.603886] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba9f189-0e67-499d-8a9d-cbed67c17fbc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.626618] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 920f5f9e-afb9-4ae6-a70b-5e902ea2c906 could not be found. [ 687.626850] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 687.627044] env[63197]: INFO nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Took 0.03 seconds to destroy the instance on the hypervisor. [ 687.627295] env[63197]: DEBUG oslo.service.loopingcall [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.627834] env[63197]: DEBUG nova.compute.manager [-] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.627933] env[63197]: DEBUG nova.network.neutron [-] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.642744] env[63197]: DEBUG nova.network.neutron [-] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.670897] env[63197]: DEBUG nova.network.neutron [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.902438] env[63197]: DEBUG nova.compute.manager [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Received event network-changed-a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.902614] env[63197]: DEBUG nova.compute.manager [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Refreshing instance network info cache due to event network-changed-a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 687.902718] env[63197]: DEBUG oslo_concurrency.lockutils [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] Acquiring lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.903061] env[63197]: DEBUG oslo_concurrency.lockutils [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] Acquired lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.903061] env[63197]: DEBUG nova.network.neutron [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Refreshing network info cache for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 687.983293] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.986s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.983962] env[63197]: ERROR nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Traceback (most recent call last): [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self.driver.spawn(context, instance, image_meta, [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] vm_ref = self.build_virtual_machine(instance, [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.983962] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] for vif in network_info: [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return self._sync_wrapper(fn, *args, **kwargs) [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self.wait() [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self[:] = self._gt.wait() [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return self._exit_event.wait() [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] result = hub.switch() [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.984272] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return self.greenlet.switch() [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] result = function(*args, **kwargs) [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] return func(*args, **kwargs) [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] raise e [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] nwinfo = self.network_api.allocate_for_instance( [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] created_port_ids = self._update_ports_for_instance( [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] with excutils.save_and_reraise_exception(): [ 687.984638] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] self.force_reraise() [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] raise self.value [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] updated_port = self._update_port( [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] _ensure_no_port_binding_failure(port) [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] raise exception.PortBindingFailed(port_id=port['id']) [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] nova.exception.PortBindingFailed: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. [ 687.984995] env[63197]: ERROR nova.compute.manager [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] [ 687.985322] env[63197]: DEBUG nova.compute.utils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.986022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.966s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.989632] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Build of instance 47da4c8f-a4cb-4d7f-81c0-7c3019169b56 was re-scheduled: Binding failed for port 3e179f8d-7685-4f62-a015-abf0b0bea999, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.989916] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.990159] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquiring lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.990507] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Acquired lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.990725] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.145404] env[63197]: DEBUG nova.network.neutron [-] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.173121] env[63197]: INFO nova.compute.manager [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] [instance: 4f2d4315-d533-431a-970a-ea655febf0ab] Took 1.03 seconds to deallocate network for instance. [ 688.419654] env[63197]: DEBUG nova.network.neutron [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.510641] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.513985] env[63197]: DEBUG nova.network.neutron [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.570898] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.648466] env[63197]: INFO nova.compute.manager [-] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Took 1.02 seconds to deallocate network for instance. [ 688.652028] env[63197]: DEBUG nova.compute.claims [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 688.652172] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.964433] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a75a4e-8070-4631-b46f-f919684c57ba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.972015] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18a64c2-460f-4822-a0ca-6feb285420a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.001851] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec642cc-dd97-407d-8bf3-c8b1ae2b6a75 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.009686] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db6255e-048e-4bc3-8054-5b5e79bbaaae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.022497] env[63197]: DEBUG oslo_concurrency.lockutils [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] Releasing lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.022733] env[63197]: DEBUG nova.compute.manager [req-e58e5f32-4146-4864-8f8e-f1d8a5d46955 req-9ed11460-c4dc-47c5-9441-dc5a02f83d53 service nova] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Received event network-vif-deleted-a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.023251] env[63197]: DEBUG nova.compute.provider_tree [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.074815] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Releasing lock "refresh_cache-47da4c8f-a4cb-4d7f-81c0-7c3019169b56" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.075039] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.075212] env[63197]: DEBUG nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.075380] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.089986] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.199744] env[63197]: INFO nova.scheduler.client.report [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Deleted allocations for instance 4f2d4315-d533-431a-970a-ea655febf0ab [ 689.529023] env[63197]: DEBUG nova.scheduler.client.report [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.593029] env[63197]: DEBUG nova.network.neutron [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.707520] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8c161916-7c3a-478d-82ef-68f4ef414cd9 tempest-ServerActionsTestOtherB-1643518160 tempest-ServerActionsTestOtherB-1643518160-project-member] Lock "4f2d4315-d533-431a-970a-ea655febf0ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.980s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.032988] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.047s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.033660] env[63197]: ERROR nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Traceback (most recent call last): [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self.driver.spawn(context, instance, image_meta, [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] vm_ref = self.build_virtual_machine(instance, [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.033660] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] for vif in network_info: [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return self._sync_wrapper(fn, *args, **kwargs) [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self.wait() [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self[:] = self._gt.wait() [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return self._exit_event.wait() [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] result = hub.switch() [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.033968] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return self.greenlet.switch() [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] result = function(*args, **kwargs) [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] return func(*args, **kwargs) [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] raise e [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] nwinfo = self.network_api.allocate_for_instance( [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] created_port_ids = self._update_ports_for_instance( [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] with excutils.save_and_reraise_exception(): [ 690.034554] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] self.force_reraise() [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] raise self.value [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] updated_port = self._update_port( [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] _ensure_no_port_binding_failure(port) [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] raise exception.PortBindingFailed(port_id=port['id']) [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] nova.exception.PortBindingFailed: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. [ 690.035089] env[63197]: ERROR nova.compute.manager [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] [ 690.035515] env[63197]: DEBUG nova.compute.utils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.035741] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.992s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.037196] env[63197]: INFO nova.compute.claims [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.040424] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Build of instance 6abf52eb-a394-477f-95e6-87cddecd4bd0 was re-scheduled: Binding failed for port ff1cf96a-19b0-402c-9174-5203beb952cf, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.040851] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.041083] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Acquiring lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.041236] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Acquired lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.041451] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.095184] env[63197]: INFO nova.compute.manager [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] [instance: 47da4c8f-a4cb-4d7f-81c0-7c3019169b56] Took 1.02 seconds to deallocate network for instance. [ 690.210156] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.559124] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.666053] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.736603] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.132360] env[63197]: INFO nova.scheduler.client.report [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Deleted allocations for instance 47da4c8f-a4cb-4d7f-81c0-7c3019169b56 [ 691.169225] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Releasing lock "refresh_cache-6abf52eb-a394-477f-95e6-87cddecd4bd0" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.169480] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.169662] env[63197]: DEBUG nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.170092] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.194528] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.433372] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46de9e4-aab5-45a2-9562-d14da7cc280f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.440997] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4f1b8e-9951-45a4-ae90-17dc65bbf7d6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.472532] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd74d30-a9c8-40b4-9fa6-707ad64050af {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.480264] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60924b7c-a06e-4834-90de-6c35d3a5497e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.493956] env[63197]: DEBUG nova.compute.provider_tree [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.642097] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8e41b2c1-9c94-4bf0-9b7d-59b28fb5290a tempest-DeleteServersAdminTestJSON-1685120198 tempest-DeleteServersAdminTestJSON-1685120198-project-member] Lock "47da4c8f-a4cb-4d7f-81c0-7c3019169b56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.878s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.698463] env[63197]: DEBUG nova.network.neutron [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.001114] env[63197]: DEBUG nova.scheduler.client.report [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.151021] env[63197]: DEBUG nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.208807] env[63197]: INFO nova.compute.manager [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] [instance: 6abf52eb-a394-477f-95e6-87cddecd4bd0] Took 1.04 seconds to deallocate network for instance. [ 692.510080] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.510080] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.511077] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.339s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.670532] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.015971] env[63197]: DEBUG nova.compute.utils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.022237] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.022237] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 693.079436] env[63197]: DEBUG nova.policy [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12d00044b65c48f484ac9fea595d6179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80138785cfdb4a2188ffb972da437646', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.236686] env[63197]: INFO nova.scheduler.client.report [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Deleted allocations for instance 6abf52eb-a394-477f-95e6-87cddecd4bd0 [ 693.444430] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Successfully created port: 11d56d25-e517-4c75-9697-b03567f0e70e {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.465102] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-499dc077-d1d3-43fa-8511-0e2cf97d0cbc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.473028] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb31b61a-23a3-4195-acbf-5bbadc6cac2f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.503461] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8150ba80-2591-4edb-8c82-78e18485b30a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.511273] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a830681d-884c-4065-b46a-becdd0e6afe1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.525441] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 693.528480] env[63197]: DEBUG nova.compute.provider_tree [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.749218] env[63197]: DEBUG oslo_concurrency.lockutils [None req-643a9386-ab4a-490c-82d9-4febc844bc60 tempest-AttachInterfacesV270Test-679842607 tempest-AttachInterfacesV270Test-679842607-project-member] Lock "6abf52eb-a394-477f-95e6-87cddecd4bd0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.470s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.038256] env[63197]: DEBUG nova.scheduler.client.report [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.217337] env[63197]: DEBUG nova.compute.manager [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Received event network-changed-11d56d25-e517-4c75-9697-b03567f0e70e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.217337] env[63197]: DEBUG nova.compute.manager [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Refreshing instance network info cache due to event network-changed-11d56d25-e517-4c75-9697-b03567f0e70e. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 694.217337] env[63197]: DEBUG oslo_concurrency.lockutils [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] Acquiring lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.217337] env[63197]: DEBUG oslo_concurrency.lockutils [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] Acquired lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.217465] env[63197]: DEBUG nova.network.neutron [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Refreshing network info cache for port 11d56d25-e517-4c75-9697-b03567f0e70e {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.252372] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.505360] env[63197]: ERROR nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. [ 694.505360] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 694.505360] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.505360] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 694.505360] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.505360] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 694.505360] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.505360] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 694.505360] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.505360] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 694.505360] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.505360] env[63197]: ERROR nova.compute.manager raise self.value [ 694.505360] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.505360] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 694.505360] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.505360] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 694.506529] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.506529] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 694.506529] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. [ 694.506529] env[63197]: ERROR nova.compute.manager [ 694.506529] env[63197]: Traceback (most recent call last): [ 694.506529] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 694.506529] env[63197]: listener.cb(fileno) [ 694.506529] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.506529] env[63197]: result = function(*args, **kwargs) [ 694.506529] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 694.506529] env[63197]: return func(*args, **kwargs) [ 694.506529] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.506529] env[63197]: raise e [ 694.506529] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.506529] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 694.506529] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.506529] env[63197]: created_port_ids = self._update_ports_for_instance( [ 694.506529] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.506529] env[63197]: with excutils.save_and_reraise_exception(): [ 694.506529] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.506529] env[63197]: self.force_reraise() [ 694.506529] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.506529] env[63197]: raise self.value [ 694.506529] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.506529] env[63197]: updated_port = self._update_port( [ 694.506529] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.506529] env[63197]: _ensure_no_port_binding_failure(port) [ 694.506529] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.506529] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 694.508682] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. [ 694.508682] env[63197]: Removing descriptor: 19 [ 694.541950] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.544628] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.034s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.545298] env[63197]: ERROR nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Traceback (most recent call last): [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self.driver.spawn(context, instance, image_meta, [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] vm_ref = self.build_virtual_machine(instance, [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.545298] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] for vif in network_info: [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return self._sync_wrapper(fn, *args, **kwargs) [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self.wait() [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self[:] = self._gt.wait() [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return self._exit_event.wait() [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] result = hub.switch() [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.545724] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return self.greenlet.switch() [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] result = function(*args, **kwargs) [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] return func(*args, **kwargs) [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] raise e [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] nwinfo = self.network_api.allocate_for_instance( [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] created_port_ids = self._update_ports_for_instance( [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] with excutils.save_and_reraise_exception(): [ 694.546124] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] self.force_reraise() [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] raise self.value [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] updated_port = self._update_port( [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] _ensure_no_port_binding_failure(port) [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] raise exception.PortBindingFailed(port_id=port['id']) [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] nova.exception.PortBindingFailed: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. [ 694.546428] env[63197]: ERROR nova.compute.manager [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] [ 694.546713] env[63197]: DEBUG nova.compute.utils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.547062] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.108s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.547238] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.547390] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 694.547660] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.830s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.549610] env[63197]: INFO nova.compute.claims [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.552277] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Build of instance 0d7e1031-896b-4ec3-9da0-74d65b1ef78a was re-scheduled: Binding failed for port 6ef74f8b-974e-4ab6-b725-a3c93b507d19, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 694.552735] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 694.552957] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Acquiring lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.553113] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Acquired lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.553318] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.558048] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e729fca-d690-4d20-8098-91c18a0d7e5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.566109] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fafe986-2b0d-4dbd-9e91-fd328488d1b0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.583571] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.583801] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.584184] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.584256] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.584408] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.584552] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.584750] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.584907] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.585104] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.585279] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.585569] env[63197]: DEBUG nova.virt.hardware [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.586251] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17aea04c-187e-4666-a54c-8ff75203e2e0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.589140] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222f191e-6614-46ef-af69-2ce40cac900b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.600036] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726291f7-d6ab-471b-80d7-d1dc319dfdd2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.602372] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22afa48e-9bc9-4f53-bb94-a2dcbfeee4d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.618661] env[63197]: ERROR nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Traceback (most recent call last): [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] yield resources [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self.driver.spawn(context, instance, image_meta, [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] vm_ref = self.build_virtual_machine(instance, [ 694.618661] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] for vif in network_info: [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] return self._sync_wrapper(fn, *args, **kwargs) [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self.wait() [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self[:] = self._gt.wait() [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] return self._exit_event.wait() [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.619049] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] current.throw(*self._exc) [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] result = function(*args, **kwargs) [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] return func(*args, **kwargs) [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] raise e [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] nwinfo = self.network_api.allocate_for_instance( [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] created_port_ids = self._update_ports_for_instance( [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] with excutils.save_and_reraise_exception(): [ 694.619393] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self.force_reraise() [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] raise self.value [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] updated_port = self._update_port( [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] _ensure_no_port_binding_failure(port) [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] raise exception.PortBindingFailed(port_id=port['id']) [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] nova.exception.PortBindingFailed: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. [ 694.619728] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] [ 694.619728] env[63197]: INFO nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Terminating instance [ 694.647128] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181485MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 694.647128] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.647128] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.746243] env[63197]: DEBUG nova.network.neutron [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.781873] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.946345] env[63197]: DEBUG nova.network.neutron [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.086260] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.220274] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.449196] env[63197]: DEBUG oslo_concurrency.lockutils [req-4d9f986f-e778-4978-8590-51728254e0cd req-aa899b14-6c43-4ec0-b361-16b7510e4e12 service nova] Releasing lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.451209] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.451209] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.725018] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Releasing lock "refresh_cache-0d7e1031-896b-4ec3-9da0-74d65b1ef78a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.725289] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 695.725463] env[63197]: DEBUG nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.725621] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.745184] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.982366] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea286ff-af5b-4499-8fbe-c56f92b8eaa8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.985539] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.994676] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f6dd18-e794-4458-abdb-9876876d4b6c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.023978] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32467f88-40b6-4e2c-9828-0edec7b68b9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.031297] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94abc6f6-794e-4716-a0b2-c21ace26db9a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.046294] env[63197]: DEBUG nova.compute.provider_tree [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.155505] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.249139] env[63197]: DEBUG nova.compute.manager [req-d35373ec-3edd-4e2e-8802-feb23a1411fe req-32e1cd1e-1433-4a6e-98b6-6848aab3693a service nova] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Received event network-vif-deleted-11d56d25-e517-4c75-9697-b03567f0e70e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.251042] env[63197]: DEBUG nova.network.neutron [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.554092] env[63197]: DEBUG nova.scheduler.client.report [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.658479] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.659536] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.659771] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.660233] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-61e18e15-b229-4f47-b281-25083767d79e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.672917] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d617cf8-1f07-4378-bce4-316ca6c6c576 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.702093] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 544ae361-619d-4e86-896c-21bf44e229d7 could not be found. [ 696.702334] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.702644] env[63197]: INFO nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 696.703110] env[63197]: DEBUG oslo.service.loopingcall [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.704022] env[63197]: DEBUG nova.compute.manager [-] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.704022] env[63197]: DEBUG nova.network.neutron [-] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.725443] env[63197]: DEBUG nova.network.neutron [-] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.755590] env[63197]: INFO nova.compute.manager [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] [instance: 0d7e1031-896b-4ec3-9da0-74d65b1ef78a] Took 1.03 seconds to deallocate network for instance. [ 697.062176] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.062176] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 697.065737] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.330s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.067997] env[63197]: INFO nova.compute.claims [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.231769] env[63197]: DEBUG nova.network.neutron [-] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.567223] env[63197]: DEBUG nova.compute.utils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.568601] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.568774] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 697.621487] env[63197]: DEBUG nova.policy [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '387a7a63289b4053b94222024d85def6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d72f64a17a2494a9d427963411170c7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.733228] env[63197]: INFO nova.compute.manager [-] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Took 1.03 seconds to deallocate network for instance. [ 697.734452] env[63197]: DEBUG nova.compute.claims [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.734452] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.812940] env[63197]: INFO nova.scheduler.client.report [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Deleted allocations for instance 0d7e1031-896b-4ec3-9da0-74d65b1ef78a [ 697.995136] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Successfully created port: 24a5e5d9-cc4d-48ee-b997-a3347522e68f {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.073531] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 698.325140] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9fd1c68a-5ee6-459b-b791-028ecdee7027 tempest-ServerActionsTestJSON-1390397092 tempest-ServerActionsTestJSON-1390397092-project-member] Lock "0d7e1031-896b-4ec3-9da0-74d65b1ef78a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.516s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.529587] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5ebadc-25e9-4111-a0c8-15a019c708da {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.537798] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1a0203-89c1-4a3d-b2da-4821f85ac549 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.571408] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7b2c38-4a2a-44a5-8bdd-7b1bbfe54546 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.578927] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbf9bad-2e58-4e57-9d31-647bf0b9a059 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.595888] env[63197]: DEBUG nova.compute.provider_tree [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.833897] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.098489] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.103151] env[63197]: DEBUG nova.scheduler.client.report [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 699.108467] env[63197]: DEBUG nova.compute.manager [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Received event network-changed-24a5e5d9-cc4d-48ee-b997-a3347522e68f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.108724] env[63197]: DEBUG nova.compute.manager [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Refreshing instance network info cache due to event network-changed-24a5e5d9-cc4d-48ee-b997-a3347522e68f. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 699.108903] env[63197]: DEBUG oslo_concurrency.lockutils [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] Acquiring lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.109047] env[63197]: DEBUG oslo_concurrency.lockutils [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] Acquired lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.109215] env[63197]: DEBUG nova.network.neutron [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Refreshing network info cache for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 699.141534] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.142424] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.142424] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.142424] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.142424] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.142424] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.142869] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.142869] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.142958] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.143110] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.143294] env[63197]: DEBUG nova.virt.hardware [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.144483] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0238b7b0-547b-40c4-92b5-3908df4a506d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.154201] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49950f9b-2ce3-4f1a-b9ce-bdc684dd8f72 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.292458] env[63197]: ERROR nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. [ 699.292458] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.292458] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.292458] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.292458] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.292458] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.292458] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.292458] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.292458] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.292458] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 699.292458] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.292458] env[63197]: ERROR nova.compute.manager raise self.value [ 699.292458] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.292458] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.292458] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.292458] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.292920] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.292920] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.292920] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. [ 699.292920] env[63197]: ERROR nova.compute.manager [ 699.292920] env[63197]: Traceback (most recent call last): [ 699.292920] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.292920] env[63197]: listener.cb(fileno) [ 699.292920] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.292920] env[63197]: result = function(*args, **kwargs) [ 699.292920] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.292920] env[63197]: return func(*args, **kwargs) [ 699.292920] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.292920] env[63197]: raise e [ 699.292920] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.292920] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 699.292920] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.292920] env[63197]: created_port_ids = self._update_ports_for_instance( [ 699.292920] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.292920] env[63197]: with excutils.save_and_reraise_exception(): [ 699.292920] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.292920] env[63197]: self.force_reraise() [ 699.292920] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.292920] env[63197]: raise self.value [ 699.292920] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.292920] env[63197]: updated_port = self._update_port( [ 699.292920] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.292920] env[63197]: _ensure_no_port_binding_failure(port) [ 699.292920] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.292920] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.293655] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. [ 699.293655] env[63197]: Removing descriptor: 19 [ 699.293655] env[63197]: ERROR nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Traceback (most recent call last): [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] yield resources [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self.driver.spawn(context, instance, image_meta, [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.293655] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] vm_ref = self.build_virtual_machine(instance, [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] for vif in network_info: [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return self._sync_wrapper(fn, *args, **kwargs) [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self.wait() [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self[:] = self._gt.wait() [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return self._exit_event.wait() [ 699.293949] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] result = hub.switch() [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return self.greenlet.switch() [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] result = function(*args, **kwargs) [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return func(*args, **kwargs) [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] raise e [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] nwinfo = self.network_api.allocate_for_instance( [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.294276] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] created_port_ids = self._update_ports_for_instance( [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] with excutils.save_and_reraise_exception(): [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self.force_reraise() [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] raise self.value [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] updated_port = self._update_port( [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] _ensure_no_port_binding_failure(port) [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.294583] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] raise exception.PortBindingFailed(port_id=port['id']) [ 699.294879] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] nova.exception.PortBindingFailed: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. [ 699.294879] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] [ 699.294879] env[63197]: INFO nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Terminating instance [ 699.301125] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Acquiring lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.359443] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.612024] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.612987] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.620054] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.723s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.654726] env[63197]: DEBUG nova.network.neutron [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.878093] env[63197]: DEBUG nova.network.neutron [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.123988] env[63197]: DEBUG nova.compute.utils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.131263] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 700.131263] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 700.262220] env[63197]: DEBUG nova.policy [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45d0ddb8ab2e463e9b90ff6fd3c16a20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcfb3f4823944481a80cc9fb92209ca7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 700.380475] env[63197]: DEBUG oslo_concurrency.lockutils [req-75eff971-1c77-4fbb-97ab-fd3e1a0b499a req-b046a7e0-42e0-442a-9241-b918fa736418 service nova] Releasing lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.380880] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Acquired lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.381076] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 700.635367] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.661318] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67d584c-bc0d-422f-9ece-f636492f9c2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.672575] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01113c8-af18-4794-921c-008f69a9a431 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.684642] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquiring lock "f8a0c645-4595-41e7-b564-2f74fed9f275" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.684928] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "f8a0c645-4595-41e7-b564-2f74fed9f275" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.715354] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494fe605-f2fc-46ec-bee4-3887781585c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.723756] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2900af9c-3080-4cbd-8941-52dbf80a9765 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.738313] env[63197]: DEBUG nova.compute.provider_tree [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.911864] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.006957] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.009260] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Successfully created port: 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.188311] env[63197]: DEBUG nova.compute.manager [req-67667dfd-438c-4a17-a274-1698b7693728 req-38a12a40-b90c-4818-aff2-401592480b7b service nova] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Received event network-vif-deleted-24a5e5d9-cc4d-48ee-b997-a3347522e68f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.241897] env[63197]: DEBUG nova.scheduler.client.report [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.513818] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Releasing lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.514268] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.514464] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 701.514765] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-147b4abf-6e73-417b-864c-d6634d93b1e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.523787] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a74039-0a33-42fc-88a7-d2eeef027db2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.549694] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7442338-8633-4866-89c4-1ef8e44004d2 could not be found. [ 701.550032] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 701.550233] env[63197]: INFO nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 701.550576] env[63197]: DEBUG oslo.service.loopingcall [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.551131] env[63197]: DEBUG nova.compute.manager [-] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.551131] env[63197]: DEBUG nova.network.neutron [-] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.570836] env[63197]: DEBUG nova.network.neutron [-] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.649669] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.693826] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:44:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='910e836c-111f-41f3-9e9f-17e7da4e1733',id=37,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1680870046',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.693826] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.693826] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.694058] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.694058] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.694058] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.694058] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.694058] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.694231] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.694231] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.694231] env[63197]: DEBUG nova.virt.hardware [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.694231] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1974b7-025d-45cb-b75f-94489b255d8c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.703355] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4550507c-5512-4552-817d-f9810eddd0bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.753049] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.134s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.753553] env[63197]: ERROR nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Traceback (most recent call last): [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self.driver.spawn(context, instance, image_meta, [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] vm_ref = self.build_virtual_machine(instance, [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.753553] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] for vif in network_info: [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return self._sync_wrapper(fn, *args, **kwargs) [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self.wait() [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self[:] = self._gt.wait() [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return self._exit_event.wait() [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] result = hub.switch() [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.753906] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return self.greenlet.switch() [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] result = function(*args, **kwargs) [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] return func(*args, **kwargs) [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] raise e [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] nwinfo = self.network_api.allocate_for_instance( [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] created_port_ids = self._update_ports_for_instance( [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] with excutils.save_and_reraise_exception(): [ 701.754244] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] self.force_reraise() [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] raise self.value [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] updated_port = self._update_port( [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] _ensure_no_port_binding_failure(port) [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] raise exception.PortBindingFailed(port_id=port['id']) [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] nova.exception.PortBindingFailed: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. [ 701.754606] env[63197]: ERROR nova.compute.manager [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] [ 701.754882] env[63197]: DEBUG nova.compute.utils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.756652] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.078s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.761518] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Build of instance 11c6fcd4-87d3-4ceb-98eb-a46ca2910584 was re-scheduled: Binding failed for port a966cac4-41f3-4e36-ba90-8c0a1c2b5765, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 701.762425] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 701.762425] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.762791] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquired lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.763352] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.073277] env[63197]: DEBUG nova.network.neutron [-] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.269550] env[63197]: ERROR nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. [ 702.269550] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 702.269550] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.269550] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 702.269550] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.269550] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 702.269550] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.269550] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 702.269550] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.269550] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 702.269550] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.269550] env[63197]: ERROR nova.compute.manager raise self.value [ 702.269550] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.269550] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 702.269550] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.269550] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 702.270034] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.270034] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 702.270034] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. [ 702.270034] env[63197]: ERROR nova.compute.manager [ 702.270034] env[63197]: Traceback (most recent call last): [ 702.270034] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 702.270034] env[63197]: listener.cb(fileno) [ 702.270034] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.270034] env[63197]: result = function(*args, **kwargs) [ 702.270034] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.270034] env[63197]: return func(*args, **kwargs) [ 702.270034] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.270034] env[63197]: raise e [ 702.270034] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.270034] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 702.270034] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.270034] env[63197]: created_port_ids = self._update_ports_for_instance( [ 702.270034] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.270034] env[63197]: with excutils.save_and_reraise_exception(): [ 702.270034] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.270034] env[63197]: self.force_reraise() [ 702.270034] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.270034] env[63197]: raise self.value [ 702.270034] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.270034] env[63197]: updated_port = self._update_port( [ 702.270034] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.270034] env[63197]: _ensure_no_port_binding_failure(port) [ 702.270034] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.270034] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 702.270807] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. [ 702.270807] env[63197]: Removing descriptor: 19 [ 702.270807] env[63197]: ERROR nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Traceback (most recent call last): [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] yield resources [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self.driver.spawn(context, instance, image_meta, [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.270807] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] vm_ref = self.build_virtual_machine(instance, [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] for vif in network_info: [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return self._sync_wrapper(fn, *args, **kwargs) [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self.wait() [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self[:] = self._gt.wait() [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return self._exit_event.wait() [ 702.271144] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] result = hub.switch() [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return self.greenlet.switch() [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] result = function(*args, **kwargs) [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return func(*args, **kwargs) [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] raise e [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] nwinfo = self.network_api.allocate_for_instance( [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.271500] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] created_port_ids = self._update_ports_for_instance( [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] with excutils.save_and_reraise_exception(): [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self.force_reraise() [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] raise self.value [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] updated_port = self._update_port( [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] _ensure_no_port_binding_failure(port) [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.271843] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] raise exception.PortBindingFailed(port_id=port['id']) [ 702.272762] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] nova.exception.PortBindingFailed: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. [ 702.272762] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] [ 702.272762] env[63197]: INFO nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Terminating instance [ 702.273270] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.273426] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquired lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.273590] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.286791] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.387354] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.577399] env[63197]: INFO nova.compute.manager [-] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Took 1.03 seconds to deallocate network for instance. [ 702.580118] env[63197]: DEBUG nova.compute.claims [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.580118] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.643021] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7806b12d-3d04-4dcb-b308-319b35e74df3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.649346] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9504a4c6-a6f8-4c83-8da9-174e55e17ec6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.680338] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771138f0-1e6d-4e51-9171-57bad9e44a96 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.690060] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6861d4-ba99-439a-89f8-f85f15b2a2cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.702229] env[63197]: DEBUG nova.compute.provider_tree [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.806226] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.876409] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.890375] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Releasing lock "refresh_cache-11c6fcd4-87d3-4ceb-98eb-a46ca2910584" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.890615] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 702.890795] env[63197]: DEBUG nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.890952] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.909022] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.205583] env[63197]: DEBUG nova.scheduler.client.report [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.212746] env[63197]: DEBUG nova.compute.manager [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Received event network-changed-10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.212855] env[63197]: DEBUG nova.compute.manager [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Refreshing instance network info cache due to event network-changed-10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 703.213048] env[63197]: DEBUG oslo_concurrency.lockutils [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] Acquiring lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.378843] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Releasing lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.379293] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.379493] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.379843] env[63197]: DEBUG oslo_concurrency.lockutils [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] Acquired lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.380035] env[63197]: DEBUG nova.network.neutron [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Refreshing network info cache for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.381145] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68e4e7d9-fa5e-404f-80d2-812015cde4df {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.392039] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9d168b-0e19-4d33-bfef-49cf70f9e036 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.414264] env[63197]: DEBUG nova.network.neutron [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.418766] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1 could not be found. [ 703.418987] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.419198] env[63197]: INFO nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 703.419437] env[63197]: DEBUG oslo.service.loopingcall [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.419647] env[63197]: DEBUG nova.compute.manager [-] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.419740] env[63197]: DEBUG nova.network.neutron [-] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.442553] env[63197]: DEBUG nova.network.neutron [-] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.713504] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.713504] env[63197]: ERROR nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. [ 703.713504] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Traceback (most recent call last): [ 703.713504] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.713504] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self.driver.spawn(context, instance, image_meta, [ 703.713504] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.713504] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.713504] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.713504] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] vm_ref = self.build_virtual_machine(instance, [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] for vif in network_info: [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return self._sync_wrapper(fn, *args, **kwargs) [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self.wait() [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self[:] = self._gt.wait() [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return self._exit_event.wait() [ 703.714284] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] result = hub.switch() [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return self.greenlet.switch() [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] result = function(*args, **kwargs) [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] return func(*args, **kwargs) [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] raise e [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] nwinfo = self.network_api.allocate_for_instance( [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.714749] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] created_port_ids = self._update_ports_for_instance( [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] with excutils.save_and_reraise_exception(): [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] self.force_reraise() [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] raise self.value [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] updated_port = self._update_port( [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] _ensure_no_port_binding_failure(port) [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.715180] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] raise exception.PortBindingFailed(port_id=port['id']) [ 703.715606] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] nova.exception.PortBindingFailed: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. [ 703.715606] env[63197]: ERROR nova.compute.manager [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] [ 703.715606] env[63197]: DEBUG nova.compute.utils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.715606] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.062s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.719724] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Build of instance 68b27a14-f52b-4d89-9cdc-be0529428eb7 was re-scheduled: Binding failed for port aafa3f61-f429-429a-ad97-fcb08a5df883, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.720247] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.720599] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Acquiring lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.720788] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Acquired lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.720954] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.903898] env[63197]: DEBUG nova.network.neutron [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.919573] env[63197]: INFO nova.compute.manager [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: 11c6fcd4-87d3-4ceb-98eb-a46ca2910584] Took 1.03 seconds to deallocate network for instance. [ 703.945585] env[63197]: DEBUG nova.network.neutron [-] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.977454] env[63197]: DEBUG nova.network.neutron [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.253311] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.349554] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.452914] env[63197]: INFO nova.compute.manager [-] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Took 1.03 seconds to deallocate network for instance. [ 704.457557] env[63197]: DEBUG nova.compute.claims [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 704.457764] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.484188] env[63197]: DEBUG oslo_concurrency.lockutils [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] Releasing lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.484188] env[63197]: DEBUG nova.compute.manager [req-7e2cd735-c5bf-477f-8d5d-e5407e3498ac req-90ee774a-c0c9-48db-b0fd-80ab65062719 service nova] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Received event network-vif-deleted-10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.661238] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2796ae-620b-4e09-9eac-734cf2846e01 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.669741] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0c60e3-efc1-48ff-8171-8628ca70cfac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.707203] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0128cda0-1d5d-44ce-8338-3c35b3d3b663 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.715622] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fa8ea6-3ab4-428f-9e13-07ba0c5d2d9a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.729056] env[63197]: DEBUG nova.compute.provider_tree [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.855290] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Releasing lock "refresh_cache-68b27a14-f52b-4d89-9cdc-be0529428eb7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.855290] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.855477] env[63197]: DEBUG nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.855605] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.871115] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.962310] env[63197]: INFO nova.scheduler.client.report [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Deleted allocations for instance 11c6fcd4-87d3-4ceb-98eb-a46ca2910584 [ 705.235550] env[63197]: DEBUG nova.scheduler.client.report [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.375413] env[63197]: DEBUG nova.network.neutron [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.475390] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b81dcbdf-8c79-4db7-99c0-87248984afd7 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "11c6fcd4-87d3-4ceb-98eb-a46ca2910584" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.672s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.542130] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Acquiring lock "fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.542351] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Lock "fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.742061] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.027s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.742726] env[63197]: ERROR nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Traceback (most recent call last): [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self.driver.spawn(context, instance, image_meta, [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] vm_ref = self.build_virtual_machine(instance, [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.742726] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] for vif in network_info: [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return self._sync_wrapper(fn, *args, **kwargs) [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self.wait() [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self[:] = self._gt.wait() [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return self._exit_event.wait() [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] result = hub.switch() [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.743107] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return self.greenlet.switch() [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] result = function(*args, **kwargs) [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] return func(*args, **kwargs) [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] raise e [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] nwinfo = self.network_api.allocate_for_instance( [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] created_port_ids = self._update_ports_for_instance( [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] with excutils.save_and_reraise_exception(): [ 705.743507] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] self.force_reraise() [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] raise self.value [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] updated_port = self._update_port( [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] _ensure_no_port_binding_failure(port) [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] raise exception.PortBindingFailed(port_id=port['id']) [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] nova.exception.PortBindingFailed: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. [ 705.744015] env[63197]: ERROR nova.compute.manager [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] [ 705.744355] env[63197]: DEBUG nova.compute.utils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.744762] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.008s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.746131] env[63197]: INFO nova.compute.claims [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.748708] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Build of instance 920f5f9e-afb9-4ae6-a70b-5e902ea2c906 was re-scheduled: Binding failed for port a4bb72f2-f8e1-4e8d-b68d-d196b43a0c04, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.749142] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.749365] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.749511] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.749664] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.877590] env[63197]: INFO nova.compute.manager [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] [instance: 68b27a14-f52b-4d89-9cdc-be0529428eb7] Took 1.02 seconds to deallocate network for instance. [ 705.976660] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 706.277597] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.386857] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.509669] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.891547] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "refresh_cache-920f5f9e-afb9-4ae6-a70b-5e902ea2c906" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.891792] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.892060] env[63197]: DEBUG nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.892265] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.903790] env[63197]: INFO nova.scheduler.client.report [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Deleted allocations for instance 68b27a14-f52b-4d89-9cdc-be0529428eb7 [ 706.911381] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.086842] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "c005d088-ee27-4d20-9db4-65b47227f22a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.087171] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "c005d088-ee27-4d20-9db4-65b47227f22a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.116659] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f826125e-c034-44e8-8972-fefce2fe13a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.125957] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a71bef5a-9786-4e50-8b8d-80732951048a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.158029] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8eb89ed-6549-4ae5-ad94-902c26571d4d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.164586] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effc719e-939b-4ea1-bc7e-fd36afba6b86 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.177590] env[63197]: DEBUG nova.compute.provider_tree [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.415797] env[63197]: DEBUG nova.network.neutron [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.418388] env[63197]: DEBUG oslo_concurrency.lockutils [None req-702e9184-f36a-467c-9325-c6d5c030f290 tempest-ServerActionsTestOtherA-1464902273 tempest-ServerActionsTestOtherA-1464902273-project-member] Lock "68b27a14-f52b-4d89-9cdc-be0529428eb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.865s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.681906] env[63197]: DEBUG nova.scheduler.client.report [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.920341] env[63197]: INFO nova.compute.manager [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 920f5f9e-afb9-4ae6-a70b-5e902ea2c906] Took 1.03 seconds to deallocate network for instance. [ 707.924740] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.187681] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.188262] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.190843] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.520s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.192570] env[63197]: INFO nova.compute.claims [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.455681] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.697362] env[63197]: DEBUG nova.compute.utils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.702019] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.702019] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 708.812031] env[63197]: DEBUG nova.policy [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6c9d629fc09245f6bd1c58ffbf989968', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d0bca9a8c2b4770bf4ea5669c80c5ee', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.965390] env[63197]: INFO nova.scheduler.client.report [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Deleted allocations for instance 920f5f9e-afb9-4ae6-a70b-5e902ea2c906 [ 709.205183] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.209785] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Successfully created port: 02e94589-6b93-48ed-913a-58f23556b583 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.474212] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7b8d9119-f9f4-4cb4-b5a4-e6c269270dd8 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "920f5f9e-afb9-4ae6-a70b-5e902ea2c906" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.682s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.627648] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2554f546-3ff6-42be-a422-036f13909246 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.635639] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d2b573-21f0-42c4-a25c-d0e34f460003 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.669754] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c4fc11-cc2d-4c89-beb2-0736c0f714bc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.677403] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24531a16-b3a8-4476-bf40-c6f80ecb5cc0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.692694] env[63197]: DEBUG nova.compute.provider_tree [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.979051] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 710.087627] env[63197]: DEBUG nova.compute.manager [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Received event network-changed-02e94589-6b93-48ed-913a-58f23556b583 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.087826] env[63197]: DEBUG nova.compute.manager [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Refreshing instance network info cache due to event network-changed-02e94589-6b93-48ed-913a-58f23556b583. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.088111] env[63197]: DEBUG oslo_concurrency.lockutils [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] Acquiring lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.088419] env[63197]: DEBUG oslo_concurrency.lockutils [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] Acquired lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.088586] env[63197]: DEBUG nova.network.neutron [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Refreshing network info cache for port 02e94589-6b93-48ed-913a-58f23556b583 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 710.200782] env[63197]: DEBUG nova.scheduler.client.report [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.214488] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.239805] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.240035] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.240197] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.240380] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.240597] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.240801] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.241197] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.241320] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.241574] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.241798] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.242036] env[63197]: DEBUG nova.virt.hardware [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.242963] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da85dfc5-be43-49ef-8554-01d3f21916f5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.250907] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4722c828-07db-4caa-b193-10214e39fcfb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.266373] env[63197]: ERROR nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. [ 710.266373] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 710.266373] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.266373] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 710.266373] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.266373] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 710.266373] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.266373] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 710.266373] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.266373] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 710.266373] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.266373] env[63197]: ERROR nova.compute.manager raise self.value [ 710.266373] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.266373] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 710.266373] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.266373] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 710.266870] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.266870] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 710.266870] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. [ 710.266870] env[63197]: ERROR nova.compute.manager [ 710.266870] env[63197]: Traceback (most recent call last): [ 710.266870] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 710.266870] env[63197]: listener.cb(fileno) [ 710.266870] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.266870] env[63197]: result = function(*args, **kwargs) [ 710.266870] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.266870] env[63197]: return func(*args, **kwargs) [ 710.266870] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.266870] env[63197]: raise e [ 710.266870] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.266870] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 710.266870] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.266870] env[63197]: created_port_ids = self._update_ports_for_instance( [ 710.266870] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.266870] env[63197]: with excutils.save_and_reraise_exception(): [ 710.266870] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.266870] env[63197]: self.force_reraise() [ 710.266870] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.266870] env[63197]: raise self.value [ 710.266870] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.266870] env[63197]: updated_port = self._update_port( [ 710.266870] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.266870] env[63197]: _ensure_no_port_binding_failure(port) [ 710.266870] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.266870] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 710.267617] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. [ 710.267617] env[63197]: Removing descriptor: 18 [ 710.267617] env[63197]: ERROR nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Traceback (most recent call last): [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] yield resources [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self.driver.spawn(context, instance, image_meta, [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.267617] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] vm_ref = self.build_virtual_machine(instance, [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] for vif in network_info: [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return self._sync_wrapper(fn, *args, **kwargs) [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self.wait() [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self[:] = self._gt.wait() [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return self._exit_event.wait() [ 710.267949] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] result = hub.switch() [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return self.greenlet.switch() [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] result = function(*args, **kwargs) [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return func(*args, **kwargs) [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] raise e [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] nwinfo = self.network_api.allocate_for_instance( [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 710.268275] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] created_port_ids = self._update_ports_for_instance( [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] with excutils.save_and_reraise_exception(): [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self.force_reraise() [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] raise self.value [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] updated_port = self._update_port( [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] _ensure_no_port_binding_failure(port) [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.268603] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] raise exception.PortBindingFailed(port_id=port['id']) [ 710.268906] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] nova.exception.PortBindingFailed: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. [ 710.268906] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] [ 710.268906] env[63197]: INFO nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Terminating instance [ 710.269790] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Acquiring lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.501121] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.604844] env[63197]: DEBUG nova.network.neutron [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.709020] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.709020] env[63197]: DEBUG nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.711901] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.065s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.838706] env[63197]: DEBUG nova.network.neutron [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.123310] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "643622a1-cb00-4829-a83a-a347e6e50e5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.123310] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "643622a1-cb00-4829-a83a-a347e6e50e5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.215598] env[63197]: DEBUG nova.compute.utils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.220251] env[63197]: DEBUG nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Not allocating networking since 'none' was specified. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 711.341881] env[63197]: DEBUG oslo_concurrency.lockutils [req-fba37a4e-7233-427d-9cb4-964c0f5d9169 req-939ee01f-b6f6-4a82-aa2f-66ba10d35c4d service nova] Releasing lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.341999] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Acquired lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.342206] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.720819] env[63197]: DEBUG nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.744253] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 544ae361-619d-4e86-896c-21bf44e229d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 711.744442] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance d7442338-8633-4866-89c4-1ef8e44004d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 711.744589] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 711.744709] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 1e650df0-b4a3-45dd-8b14-b41d4c77b625 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 711.744823] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 09dc9549-ab9a-4f36-9671-ab1b5595649d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 711.867036] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.942793] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.111298] env[63197]: DEBUG nova.compute.manager [req-7b34f898-2ab7-4afa-9e44-ea2036f7e1b4 req-448f1ea5-3504-4094-b363-f5f7aca8ba12 service nova] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Received event network-vif-deleted-02e94589-6b93-48ed-913a-58f23556b583 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 712.247406] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 0f566f8c-c92a-4885-94a6-08043857a1c2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.445140] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Releasing lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.445575] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 712.445772] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.446090] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44a612ba-8881-49ff-930b-ff22657a50ba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.454832] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccde2eb-51ee-4403-83ea-85db46e97d3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.475339] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1e650df0-b4a3-45dd-8b14-b41d4c77b625 could not be found. [ 712.475563] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 712.477413] env[63197]: INFO nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Took 0.03 seconds to destroy the instance on the hypervisor. [ 712.477413] env[63197]: DEBUG oslo.service.loopingcall [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.477413] env[63197]: DEBUG nova.compute.manager [-] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.477413] env[63197]: DEBUG nova.network.neutron [-] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.492371] env[63197]: DEBUG nova.network.neutron [-] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.732593] env[63197]: DEBUG nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.751574] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a0638d5a-6d33-4eb6-a9b6-ad6620219259 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 712.758300] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.758577] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.758734] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.758914] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.759071] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.759218] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.759417] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.759571] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.759728] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.759884] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.760063] env[63197]: DEBUG nova.virt.hardware [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.761185] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f71f740-9a7f-4b3e-b3ed-d9dea082cee0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.770013] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565f2368-2737-4d9c-8a69-686999e2d3f9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.783129] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 712.788530] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Creating folder: Project (5e3e59ef304b43e1982ffca0e40c15e5). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 712.788772] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8421d46-2307-4511-85f3-db720bac59e3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.798087] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Created folder: Project (5e3e59ef304b43e1982ffca0e40c15e5) in parent group-v290286. [ 712.798266] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Creating folder: Instances. Parent ref: group-v290303. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 712.798475] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3d5fc54-dcba-4a06-8a04-507ac035ac55 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.807670] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Created folder: Instances in parent group-v290303. [ 712.807884] env[63197]: DEBUG oslo.service.loopingcall [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 712.808076] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 712.808262] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bec2642-18b3-4596-b600-411c53a143e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.824179] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 712.824179] env[63197]: value = "task-1364085" [ 712.824179] env[63197]: _type = "Task" [ 712.824179] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.830882] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364085, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.995368] env[63197]: DEBUG nova.network.neutron [-] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.254797] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 715c53a9-224e-4457-9df9-a2ada9fcae93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 713.334573] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364085, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.497808] env[63197]: INFO nova.compute.manager [-] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Took 1.02 seconds to deallocate network for instance. [ 713.500389] env[63197]: DEBUG nova.compute.claims [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 713.500643] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.757605] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 86a728c9-27d3-4c98-a4d6-6b0736f24648 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 713.835138] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364085, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.260738] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 218ad539-e2d0-4d4b-9857-ec9d2af0be03 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 714.336025] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364085, 'name': CreateVM_Task, 'duration_secs': 1.250436} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.336025] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 714.336269] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.336425] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.336754] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 714.337011] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d862983d-a88b-411c-8046-d74f51e1a96a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.342200] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 714.342200] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527ecd55-8e9d-fcde-8e12-acc37eda5dea" [ 714.342200] env[63197]: _type = "Task" [ 714.342200] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.350491] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527ecd55-8e9d-fcde-8e12-acc37eda5dea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.765332] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance d9f1dcc1-77fd-4c8c-b930-eb421691dc94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 714.851865] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527ecd55-8e9d-fcde-8e12-acc37eda5dea, 'name': SearchDatastore_Task, 'duration_secs': 0.011548} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.852193] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.852424] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 714.852654] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.852811] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.852974] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 714.853235] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3e28927-6304-4fd5-b992-53fdf8517c78 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.861052] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 714.861232] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 714.861946] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1378489a-f0bf-44d7-8f46-c2ea3a89b336 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.866696] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 714.866696] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522d8c31-dbae-e786-b435-7967c7db29b7" [ 714.866696] env[63197]: _type = "Task" [ 714.866696] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.874284] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522d8c31-dbae-e786-b435-7967c7db29b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.268625] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a1ba5772-9e73-4a51-ae62-d9c17a64ff0d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 715.377222] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522d8c31-dbae-e786-b435-7967c7db29b7, 'name': SearchDatastore_Task, 'duration_secs': 0.009676} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.378366] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-697b9b18-1573-4e24-bcf0-41b8f7383a40 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.383157] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 715.383157] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280f285-a3ab-0697-0342-00b7012b4a3a" [ 715.383157] env[63197]: _type = "Task" [ 715.383157] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.391138] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280f285-a3ab-0697-0342-00b7012b4a3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.771752] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance e28f14fc-dd09-4baf-9be9-cba280f47f80 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 715.893336] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280f285-a3ab-0697-0342-00b7012b4a3a, 'name': SearchDatastore_Task, 'duration_secs': 0.008769} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.893609] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.893862] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 09dc9549-ab9a-4f36-9671-ab1b5595649d/09dc9549-ab9a-4f36-9671-ab1b5595649d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 715.894121] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40ef6e02-e7a6-4cef-ae57-623c5e5bc18f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.901965] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 715.901965] env[63197]: value = "task-1364086" [ 715.901965] env[63197]: _type = "Task" [ 715.901965] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.911015] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364086, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.276052] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 6c3137a4-4e51-4bf9-b7d3-779de61b69b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 716.411982] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364086, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457477} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.412274] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 09dc9549-ab9a-4f36-9671-ab1b5595649d/09dc9549-ab9a-4f36-9671-ab1b5595649d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 716.412482] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 716.412727] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4aeb2599-3bb9-4b22-ae7e-eff46a2581e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.419190] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 716.419190] env[63197]: value = "task-1364087" [ 716.419190] env[63197]: _type = "Task" [ 716.419190] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.427298] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364087, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.779200] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 15120438-f2a8-4295-a70f-6ec8afdd9030 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 716.929027] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364087, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057851} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.929027] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 716.929736] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c862549b-1b8b-466f-9750-0b13ff2b2fa2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.948626] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 09dc9549-ab9a-4f36-9671-ab1b5595649d/09dc9549-ab9a-4f36-9671-ab1b5595649d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 716.948876] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5aeb22b6-e66f-44b2-af5b-1ec4fc03aa99 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.968176] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 716.968176] env[63197]: value = "task-1364088" [ 716.968176] env[63197]: _type = "Task" [ 716.968176] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.975548] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364088, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.283071] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance b5935562-9a77-4b92-ac73-a7af496ccced has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 717.478019] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364088, 'name': ReconfigVM_Task, 'duration_secs': 0.280887} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.478305] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 09dc9549-ab9a-4f36-9671-ab1b5595649d/09dc9549-ab9a-4f36-9671-ab1b5595649d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 717.478997] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e800943-d397-4109-b695-d519efd52953 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.485386] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 717.485386] env[63197]: value = "task-1364089" [ 717.485386] env[63197]: _type = "Task" [ 717.485386] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.492827] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364089, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.785700] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a90e6999-9aab-4b83-98e4-2b71681d1bb5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 717.995447] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364089, 'name': Rename_Task, 'duration_secs': 0.125527} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.995736] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 717.996352] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95ce0647-0338-4953-8375-61bd7907d9ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.002778] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 718.002778] env[63197]: value = "task-1364090" [ 718.002778] env[63197]: _type = "Task" [ 718.002778] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.010307] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364090, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.288578] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance bd771a95-98b1-4fdb-b213-42d693e027f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 718.514690] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364090, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.791557] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance d7a2de2b-d14b-437c-93ff-17ee24bd97ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 719.013146] env[63197]: DEBUG oslo_vmware.api [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364090, 'name': PowerOnVM_Task, 'duration_secs': 0.569285} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.013405] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 719.013600] env[63197]: INFO nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Took 6.28 seconds to spawn the instance on the hypervisor. [ 719.013775] env[63197]: DEBUG nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 719.014551] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6dbe69b-834e-4e99-8cbf-28be26ed1d57 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.294849] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 82919ffb-2b50-4336-9517-c741d259f19e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 719.538114] env[63197]: INFO nova.compute.manager [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Took 26.89 seconds to build instance. [ 719.797334] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 66af0137-22d4-4670-8225-8d9b9dd2884a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 720.040575] env[63197]: DEBUG nova.compute.manager [None req-75de1ea7-d74d-42cc-a7bb-bef8ca35743c tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 720.041592] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4a3b33-92c5-4494-949d-3b7881026e7e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.044486] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5dc10443-78fe-4180-ab2d-a394a4deb747 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "09dc9549-ab9a-4f36-9671-ab1b5595649d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.234s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.144045] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "09dc9549-ab9a-4f36-9671-ab1b5595649d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.144045] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "09dc9549-ab9a-4f36-9671-ab1b5595649d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.144045] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "09dc9549-ab9a-4f36-9671-ab1b5595649d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.144045] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "09dc9549-ab9a-4f36-9671-ab1b5595649d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.144252] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "09dc9549-ab9a-4f36-9671-ab1b5595649d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.146505] env[63197]: INFO nova.compute.manager [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Terminating instance [ 720.148079] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "refresh_cache-09dc9549-ab9a-4f36-9671-ab1b5595649d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.148236] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquired lock "refresh_cache-09dc9549-ab9a-4f36-9671-ab1b5595649d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.148424] env[63197]: DEBUG nova.network.neutron [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.299844] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 720.546751] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.552421] env[63197]: INFO nova.compute.manager [None req-75de1ea7-d74d-42cc-a7bb-bef8ca35743c tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] instance snapshotting [ 720.552970] env[63197]: DEBUG nova.objects.instance [None req-75de1ea7-d74d-42cc-a7bb-bef8ca35743c tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lazy-loading 'flavor' on Instance uuid 09dc9549-ab9a-4f36-9671-ab1b5595649d {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 720.665085] env[63197]: DEBUG nova.network.neutron [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.725548] env[63197]: DEBUG nova.network.neutron [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.802579] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 721.060837] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be18dd0-5d3d-4f74-976c-256ab2e9d381 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.079305] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.080145] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc8ef6b-7296-4675-ac8e-7bd0039f358c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.228795] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Releasing lock "refresh_cache-09dc9549-ab9a-4f36-9671-ab1b5595649d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.229194] env[63197]: DEBUG nova.compute.manager [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 721.229393] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.230297] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00e062f-003c-4a2e-ac6f-4271443245b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.238248] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 721.238511] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6303d26d-b8a5-49a4-a0c4-7bcbbe338d45 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.244244] env[63197]: DEBUG oslo_vmware.api [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 721.244244] env[63197]: value = "task-1364091" [ 721.244244] env[63197]: _type = "Task" [ 721.244244] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.252623] env[63197]: DEBUG oslo_vmware.api [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364091, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.305770] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 7786e5f4-a7c9-4c14-b407-d2eda85bf37a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 721.591843] env[63197]: DEBUG nova.compute.manager [None req-75de1ea7-d74d-42cc-a7bb-bef8ca35743c tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Instance disappeared during snapshot {{(pid=63197) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 721.735993] env[63197]: DEBUG nova.compute.manager [None req-75de1ea7-d74d-42cc-a7bb-bef8ca35743c tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Found 0 images (rotation: 2) {{(pid=63197) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 721.756313] env[63197]: DEBUG oslo_vmware.api [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364091, 'name': PowerOffVM_Task, 'duration_secs': 0.184879} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.757361] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 721.757674] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 721.757876] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e0ea67c-b728-4bf6-932e-41dddaf1a9b0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.783510] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 721.783811] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 721.784038] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Deleting the datastore file [datastore2] 09dc9549-ab9a-4f36-9671-ab1b5595649d {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 721.784336] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0876b63e-692e-43bc-8109-e2ee1a2dffce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.791085] env[63197]: DEBUG oslo_vmware.api [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for the task: (returnval){ [ 721.791085] env[63197]: value = "task-1364093" [ 721.791085] env[63197]: _type = "Task" [ 721.791085] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.799237] env[63197]: DEBUG oslo_vmware.api [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.809097] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 5128f0c1-da4a-4eb4-9dde-23f830800512 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 722.301016] env[63197]: DEBUG oslo_vmware.api [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Task: {'id': task-1364093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094284} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.301270] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 722.301448] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 722.301803] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.302012] env[63197]: INFO nova.compute.manager [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Took 1.07 seconds to destroy the instance on the hypervisor. [ 722.302290] env[63197]: DEBUG oslo.service.loopingcall [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.302501] env[63197]: DEBUG nova.compute.manager [-] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.302600] env[63197]: DEBUG nova.network.neutron [-] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.312092] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 06167654-0ac8-4b04-8eb7-071eac0894ae has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 722.319183] env[63197]: DEBUG nova.network.neutron [-] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.815558] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance f8a0c645-4595-41e7-b564-2f74fed9f275 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 722.821537] env[63197]: DEBUG nova.network.neutron [-] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.319132] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance fc4d4223-b0cb-4a40-bd5f-62ec740b57d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 723.327329] env[63197]: INFO nova.compute.manager [-] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Took 1.02 seconds to deallocate network for instance. [ 723.822746] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance c005d088-ee27-4d20-9db4-65b47227f22a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 723.833904] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.326114] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 643622a1-cb00-4829-a83a-a347e6e50e5d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 724.326434] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 724.326519] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 724.631234] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f54bb60-8ec2-44e7-8a7d-b286043b16e9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.639364] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4fc7ed-8568-4a0a-a23e-45aee28da77a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.672228] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d35bdb-93dd-4b9d-a64b-cdafbf51344e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.680049] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa73ad86-a4cb-4568-a605-d72c69b59afb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.694266] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.198157] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.704107] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 725.704107] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.992s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.704107] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.922s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.705531] env[63197]: INFO nova.compute.claims [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.031991] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08c2432-c4c5-4573-a104-b26c191f2aea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.039738] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc678bd-dc9d-421e-bb03-ad6383bbee3a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.068370] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4956c349-9d29-4974-b500-41403355a544 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.075430] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe49a85-e7f2-4f7c-a1c9-f9ea64bc18a3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.088196] env[63197]: DEBUG nova.compute.provider_tree [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.591052] env[63197]: DEBUG nova.scheduler.client.report [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.095877] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.096437] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.099241] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.365s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.605730] env[63197]: DEBUG nova.compute.utils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.609997] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.610182] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.662371] env[63197]: DEBUG nova.policy [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31a519690def4b6b92028150098f7801', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad6294feb57a407d9316ef6bc861262b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.936775] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df27e70e-aacf-4c2a-8a9e-9bfe9d3c76e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.941950] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Successfully created port: fc86d49c-c9ff-4aef-badc-b94cd8d432bf {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.947048] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb5dad0-f6e3-453e-ad72-60a0d7baba21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.976062] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f110cd-92ea-4166-bdfa-596d7cc5e6e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.983232] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d8144d-325c-43ef-a05e-d2e815837a29 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.995945] env[63197]: DEBUG nova.compute.provider_tree [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.111195] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.500018] env[63197]: DEBUG nova.scheduler.client.report [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.721414] env[63197]: DEBUG nova.compute.manager [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Received event network-changed-fc86d49c-c9ff-4aef-badc-b94cd8d432bf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.721608] env[63197]: DEBUG nova.compute.manager [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Refreshing instance network info cache due to event network-changed-fc86d49c-c9ff-4aef-badc-b94cd8d432bf. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.721839] env[63197]: DEBUG oslo_concurrency.lockutils [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] Acquiring lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.721980] env[63197]: DEBUG oslo_concurrency.lockutils [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] Acquired lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.722150] env[63197]: DEBUG nova.network.neutron [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Refreshing network info cache for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.915975] env[63197]: ERROR nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. [ 729.915975] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.915975] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.915975] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.915975] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.915975] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.915975] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.915975] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.915975] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.915975] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 729.915975] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.915975] env[63197]: ERROR nova.compute.manager raise self.value [ 729.915975] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.915975] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.915975] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.915975] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.916458] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.916458] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.916458] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. [ 729.916458] env[63197]: ERROR nova.compute.manager [ 729.916458] env[63197]: Traceback (most recent call last): [ 729.916458] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.916458] env[63197]: listener.cb(fileno) [ 729.916458] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.916458] env[63197]: result = function(*args, **kwargs) [ 729.916458] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.916458] env[63197]: return func(*args, **kwargs) [ 729.916458] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.916458] env[63197]: raise e [ 729.916458] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.916458] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 729.916458] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.916458] env[63197]: created_port_ids = self._update_ports_for_instance( [ 729.916458] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.916458] env[63197]: with excutils.save_and_reraise_exception(): [ 729.916458] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.916458] env[63197]: self.force_reraise() [ 729.916458] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.916458] env[63197]: raise self.value [ 729.916458] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.916458] env[63197]: updated_port = self._update_port( [ 729.916458] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.916458] env[63197]: _ensure_no_port_binding_failure(port) [ 729.916458] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.916458] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.917200] env[63197]: nova.exception.PortBindingFailed: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. [ 729.917200] env[63197]: Removing descriptor: 18 [ 730.005217] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.005892] env[63197]: ERROR nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Traceback (most recent call last): [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self.driver.spawn(context, instance, image_meta, [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] vm_ref = self.build_virtual_machine(instance, [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.005892] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] for vif in network_info: [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] return self._sync_wrapper(fn, *args, **kwargs) [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self.wait() [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self[:] = self._gt.wait() [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] return self._exit_event.wait() [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] current.throw(*self._exc) [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.006233] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] result = function(*args, **kwargs) [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] return func(*args, **kwargs) [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] raise e [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] nwinfo = self.network_api.allocate_for_instance( [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] created_port_ids = self._update_ports_for_instance( [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] with excutils.save_and_reraise_exception(): [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] self.force_reraise() [ 730.006546] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] raise self.value [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] updated_port = self._update_port( [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] _ensure_no_port_binding_failure(port) [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] raise exception.PortBindingFailed(port_id=port['id']) [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] nova.exception.PortBindingFailed: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. [ 730.006865] env[63197]: ERROR nova.compute.manager [instance: 544ae361-619d-4e86-896c-21bf44e229d7] [ 730.006865] env[63197]: DEBUG nova.compute.utils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.007864] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.649s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.009967] env[63197]: INFO nova.compute.claims [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.013126] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Build of instance 544ae361-619d-4e86-896c-21bf44e229d7 was re-scheduled: Binding failed for port 11d56d25-e517-4c75-9697-b03567f0e70e, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 730.014030] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 730.014030] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.014030] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.014030] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.119935] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 730.144388] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.144636] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.144792] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.144993] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.145129] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.145313] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.145945] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.146160] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.146669] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.146669] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.146669] env[63197]: DEBUG nova.virt.hardware [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.147534] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce984dd1-ea40-4dc5-a30d-ce56639495d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.155775] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26cd0102-314b-4363-9010-f6f60cc7b89e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.170910] env[63197]: ERROR nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Traceback (most recent call last): [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] yield resources [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self.driver.spawn(context, instance, image_meta, [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] vm_ref = self.build_virtual_machine(instance, [ 730.170910] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] for vif in network_info: [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] return self._sync_wrapper(fn, *args, **kwargs) [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self.wait() [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self[:] = self._gt.wait() [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] return self._exit_event.wait() [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.171308] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] current.throw(*self._exc) [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] result = function(*args, **kwargs) [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] return func(*args, **kwargs) [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] raise e [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] nwinfo = self.network_api.allocate_for_instance( [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] created_port_ids = self._update_ports_for_instance( [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] with excutils.save_and_reraise_exception(): [ 730.171718] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self.force_reraise() [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] raise self.value [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] updated_port = self._update_port( [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] _ensure_no_port_binding_failure(port) [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] raise exception.PortBindingFailed(port_id=port['id']) [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] nova.exception.PortBindingFailed: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. [ 730.172112] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] [ 730.172112] env[63197]: INFO nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Terminating instance [ 730.173304] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.243625] env[63197]: DEBUG nova.network.neutron [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.319982] env[63197]: DEBUG nova.network.neutron [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.533744] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.625387] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.822575] env[63197]: DEBUG oslo_concurrency.lockutils [req-357f9fd9-4a49-471c-9d2a-30d09aa40198 req-685624a6-1398-4802-a2e3-f070c2f0beb5 service nova] Releasing lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.823099] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquired lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.823350] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 731.127697] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-544ae361-619d-4e86-896c-21bf44e229d7" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.127996] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.128233] env[63197]: DEBUG nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.128469] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.145201] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.342198] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.393375] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6f33d3-f595-4dba-a383-2ce5512a1459 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.403082] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac9e1fa-ab39-461f-9265-76e8764461bd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.434087] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.435632] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00d1e53-a3f1-4ae8-a95c-9b50725ac2c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.443593] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bf4fe8-d892-4cd0-a623-6361ee9adfec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.457130] env[63197]: DEBUG nova.compute.provider_tree [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.652033] env[63197]: DEBUG nova.network.neutron [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.748318] env[63197]: DEBUG nova.compute.manager [req-2f0d93ba-ff94-4f82-bf7f-231be35f4544 req-b5062a05-09c2-4725-848f-c9d779d22a2c service nova] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Received event network-vif-deleted-fc86d49c-c9ff-4aef-badc-b94cd8d432bf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.938927] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Releasing lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.939390] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 731.939623] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.939932] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c0c7166-59fa-4af8-b508-ccaa7a782e21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.949502] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a7ef82-e0c0-49a3-9263-a785ba6cc91b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.959719] env[63197]: DEBUG nova.scheduler.client.report [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.974035] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f566f8c-c92a-4885-94a6-08043857a1c2 could not be found. [ 731.974035] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 731.974208] env[63197]: INFO nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 731.974832] env[63197]: DEBUG oslo.service.loopingcall [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.974832] env[63197]: DEBUG nova.compute.manager [-] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.974832] env[63197]: DEBUG nova.network.neutron [-] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.989749] env[63197]: DEBUG nova.network.neutron [-] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.156599] env[63197]: INFO nova.compute.manager [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 544ae361-619d-4e86-896c-21bf44e229d7] Took 1.03 seconds to deallocate network for instance. [ 732.464625] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.465166] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 732.467825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.888s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.493898] env[63197]: DEBUG nova.network.neutron [-] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.972481] env[63197]: DEBUG nova.compute.utils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.977109] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.977506] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 732.996700] env[63197]: INFO nova.compute.manager [-] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Took 1.02 seconds to deallocate network for instance. [ 732.998947] env[63197]: DEBUG nova.compute.claims [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 732.999139] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.019284] env[63197]: DEBUG nova.policy [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 733.190173] env[63197]: INFO nova.scheduler.client.report [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted allocations for instance 544ae361-619d-4e86-896c-21bf44e229d7 [ 733.384044] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0fa9c1-2270-42cf-b06a-3cead5efc56a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.393723] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a67f83b-078c-436d-bc1d-1e0216d0732a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.425861] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e31922-9d35-487a-9d42-0b8283dccf6a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.433413] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baed9d20-e4d8-4470-867c-044a9eaea7a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.447460] env[63197]: DEBUG nova.compute.provider_tree [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.477641] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 733.499489] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Successfully created port: 35397525-7932-47df-a882-adda1487c79b {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 733.702226] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b597cafa-3d43-42d3-a820-723152bc7a2e tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "544ae361-619d-4e86-896c-21bf44e229d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.245s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.951175] env[63197]: DEBUG nova.scheduler.client.report [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.203038] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.439851] env[63197]: DEBUG nova.compute.manager [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Received event network-changed-35397525-7932-47df-a882-adda1487c79b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.440082] env[63197]: DEBUG nova.compute.manager [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Refreshing instance network info cache due to event network-changed-35397525-7932-47df-a882-adda1487c79b. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 734.440301] env[63197]: DEBUG oslo_concurrency.lockutils [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] Acquiring lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.440443] env[63197]: DEBUG oslo_concurrency.lockutils [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] Acquired lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.440665] env[63197]: DEBUG nova.network.neutron [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Refreshing network info cache for port 35397525-7932-47df-a882-adda1487c79b {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 734.456260] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.456877] env[63197]: ERROR nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Traceback (most recent call last): [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self.driver.spawn(context, instance, image_meta, [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] vm_ref = self.build_virtual_machine(instance, [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.456877] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] for vif in network_info: [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return self._sync_wrapper(fn, *args, **kwargs) [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self.wait() [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self[:] = self._gt.wait() [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return self._exit_event.wait() [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] result = hub.switch() [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.457244] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return self.greenlet.switch() [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] result = function(*args, **kwargs) [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] return func(*args, **kwargs) [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] raise e [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] nwinfo = self.network_api.allocate_for_instance( [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] created_port_ids = self._update_ports_for_instance( [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] with excutils.save_and_reraise_exception(): [ 734.457565] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] self.force_reraise() [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] raise self.value [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] updated_port = self._update_port( [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] _ensure_no_port_binding_failure(port) [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] raise exception.PortBindingFailed(port_id=port['id']) [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] nova.exception.PortBindingFailed: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. [ 734.457889] env[63197]: ERROR nova.compute.manager [instance: d7442338-8633-4866-89c4-1ef8e44004d2] [ 734.458181] env[63197]: DEBUG nova.compute.utils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.459312] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.002s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.462613] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Build of instance d7442338-8633-4866-89c4-1ef8e44004d2 was re-scheduled: Binding failed for port 24a5e5d9-cc4d-48ee-b997-a3347522e68f, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 734.464057] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 734.464307] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Acquiring lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.464462] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Acquired lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.464623] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.491532] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.519635] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.519900] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.520054] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.520238] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.520381] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.520525] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.520756] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.520911] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.521725] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.521972] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.522184] env[63197]: DEBUG nova.virt.hardware [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.523065] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee24e02-24a0-45a5-845f-bf4d3fa45735 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.531400] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b6282a-bd1b-475d-8209-8753ab08fbe1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.662614] env[63197]: ERROR nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. [ 734.662614] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 734.662614] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.662614] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 734.662614] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.662614] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 734.662614] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.662614] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 734.662614] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.662614] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 734.662614] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.662614] env[63197]: ERROR nova.compute.manager raise self.value [ 734.662614] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.662614] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 734.662614] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.662614] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 734.663193] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.663193] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 734.663193] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. [ 734.663193] env[63197]: ERROR nova.compute.manager [ 734.663193] env[63197]: Traceback (most recent call last): [ 734.663193] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 734.663193] env[63197]: listener.cb(fileno) [ 734.663193] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.663193] env[63197]: result = function(*args, **kwargs) [ 734.663193] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.663193] env[63197]: return func(*args, **kwargs) [ 734.663193] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.663193] env[63197]: raise e [ 734.663193] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.663193] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 734.663193] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.663193] env[63197]: created_port_ids = self._update_ports_for_instance( [ 734.663193] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.663193] env[63197]: with excutils.save_and_reraise_exception(): [ 734.663193] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.663193] env[63197]: self.force_reraise() [ 734.663193] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.663193] env[63197]: raise self.value [ 734.663193] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.663193] env[63197]: updated_port = self._update_port( [ 734.663193] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.663193] env[63197]: _ensure_no_port_binding_failure(port) [ 734.663193] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.663193] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 734.664062] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. [ 734.664062] env[63197]: Removing descriptor: 18 [ 734.664062] env[63197]: ERROR nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Traceback (most recent call last): [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] yield resources [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self.driver.spawn(context, instance, image_meta, [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.664062] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] vm_ref = self.build_virtual_machine(instance, [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] for vif in network_info: [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return self._sync_wrapper(fn, *args, **kwargs) [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self.wait() [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self[:] = self._gt.wait() [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return self._exit_event.wait() [ 734.664439] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] result = hub.switch() [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return self.greenlet.switch() [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] result = function(*args, **kwargs) [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return func(*args, **kwargs) [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] raise e [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] nwinfo = self.network_api.allocate_for_instance( [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.664971] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] created_port_ids = self._update_ports_for_instance( [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] with excutils.save_and_reraise_exception(): [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self.force_reraise() [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] raise self.value [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] updated_port = self._update_port( [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] _ensure_no_port_binding_failure(port) [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.665711] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] raise exception.PortBindingFailed(port_id=port['id']) [ 734.667124] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] nova.exception.PortBindingFailed: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. [ 734.667124] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] [ 734.667124] env[63197]: INFO nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Terminating instance [ 734.667753] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.723834] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.959139] env[63197]: DEBUG nova.network.neutron [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.990810] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.045737] env[63197]: DEBUG nova.network.neutron [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.109855] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.396812] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7929fa7b-8cb7-4a10-9790-cfd7ff2eefe4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.404775] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1c87ea-7c42-4703-bcab-fe3784003c4b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.434956] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d1e5b2-c868-4a35-8d9f-03cecf7854c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.441840] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e53a32a-eec3-45f7-a23a-efd1a2fc4f2c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.454194] env[63197]: DEBUG nova.compute.provider_tree [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.551122] env[63197]: DEBUG oslo_concurrency.lockutils [req-4aa6b6eb-33ca-426b-b745-a568f081f203 req-44a516d9-fe37-4fc2-a486-7f16c476422b service nova] Releasing lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.551472] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.551659] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.609793] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "92ee8e7a-c169-44d8-8f74-84590b4f62e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.609793] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "92ee8e7a-c169-44d8-8f74-84590b4f62e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.614604] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Releasing lock "refresh_cache-d7442338-8633-4866-89c4-1ef8e44004d2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.614604] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 735.614604] env[63197]: DEBUG nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 735.614604] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.637638] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.957264] env[63197]: DEBUG nova.scheduler.client.report [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.070524] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.140361] env[63197]: DEBUG nova.network.neutron [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.158380] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.462202] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.462845] env[63197]: ERROR nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Traceback (most recent call last): [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self.driver.spawn(context, instance, image_meta, [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] vm_ref = self.build_virtual_machine(instance, [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.462845] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] for vif in network_info: [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return self._sync_wrapper(fn, *args, **kwargs) [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self.wait() [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self[:] = self._gt.wait() [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return self._exit_event.wait() [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] result = hub.switch() [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 736.463188] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return self.greenlet.switch() [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] result = function(*args, **kwargs) [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] return func(*args, **kwargs) [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] raise e [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] nwinfo = self.network_api.allocate_for_instance( [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] created_port_ids = self._update_ports_for_instance( [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] with excutils.save_and_reraise_exception(): [ 736.463633] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] self.force_reraise() [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] raise self.value [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] updated_port = self._update_port( [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] _ensure_no_port_binding_failure(port) [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] raise exception.PortBindingFailed(port_id=port['id']) [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] nova.exception.PortBindingFailed: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. [ 736.464035] env[63197]: ERROR nova.compute.manager [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] [ 736.464363] env[63197]: DEBUG nova.compute.utils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.464735] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.956s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.466153] env[63197]: INFO nova.compute.claims [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.468790] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Build of instance 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1 was re-scheduled: Binding failed for port 10eba3f5-6de1-4e63-b6af-1e8aa1bdb8ad, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 736.469246] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 736.469466] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquiring lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.469608] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Acquired lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.469760] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.472016] env[63197]: DEBUG nova.compute.manager [req-6e6f5220-1d76-4255-a5e0-bba1c942ffe6 req-ffb7438d-d428-4164-86d5-45e9758620ed service nova] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Received event network-vif-deleted-35397525-7932-47df-a882-adda1487c79b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.644693] env[63197]: INFO nova.compute.manager [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] [instance: d7442338-8633-4866-89c4-1ef8e44004d2] Took 1.03 seconds to deallocate network for instance. [ 736.661524] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.661990] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 736.662201] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 736.662489] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-411500d4-86c4-4433-a259-21969b81aa23 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.671878] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9322fe6b-9e9b-498a-9a78-f033124524a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.694416] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a0638d5a-6d33-4eb6-a9b6-ad6620219259 could not be found. [ 736.694637] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 736.694819] env[63197]: INFO nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Took 0.03 seconds to destroy the instance on the hypervisor. [ 736.695072] env[63197]: DEBUG oslo.service.loopingcall [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.695296] env[63197]: DEBUG nova.compute.manager [-] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.695389] env[63197]: DEBUG nova.network.neutron [-] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.709993] env[63197]: DEBUG nova.network.neutron [-] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.990960] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.050220] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.213883] env[63197]: DEBUG nova.network.neutron [-] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.553040] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Releasing lock "refresh_cache-7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.553324] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 737.553457] env[63197]: DEBUG nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.553622] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.569045] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.679315] env[63197]: INFO nova.scheduler.client.report [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Deleted allocations for instance d7442338-8633-4866-89c4-1ef8e44004d2 [ 737.716393] env[63197]: INFO nova.compute.manager [-] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Took 1.02 seconds to deallocate network for instance. [ 737.720220] env[63197]: DEBUG nova.compute.claims [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 737.720420] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.873080] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e902b62f-aac3-491b-811a-2028109b9b69 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.881150] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5fe1de-3d2e-4144-96f8-c645f7f2b902 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.911207] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2694cbd-e31e-47bf-b6a8-3a33ec99b201 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.918675] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b65676-de78-46f5-a122-7a15205a6df9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.932248] env[63197]: DEBUG nova.compute.provider_tree [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.071904] env[63197]: DEBUG nova.network.neutron [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.195322] env[63197]: DEBUG oslo_concurrency.lockutils [None req-83783282-f81d-4f70-b45a-9941e78f71d7 tempest-ServerAddressesNegativeTestJSON-338251652 tempest-ServerAddressesNegativeTestJSON-338251652-project-member] Lock "d7442338-8633-4866-89c4-1ef8e44004d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.890s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.435479] env[63197]: DEBUG nova.scheduler.client.report [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.576012] env[63197]: INFO nova.compute.manager [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] [instance: 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1] Took 1.02 seconds to deallocate network for instance. [ 738.699395] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 738.941063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.941880] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.945122] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.490s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.946974] env[63197]: INFO nova.compute.claims [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.228530] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.455412] env[63197]: DEBUG nova.compute.utils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.458437] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.458629] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 739.513376] env[63197]: DEBUG nova.policy [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '40f23bcb7b2f4040b07a9a86a0c2a992', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3c12badef554b8e8bdf2fd8c255c422', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.610122] env[63197]: INFO nova.scheduler.client.report [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Deleted allocations for instance 7b2f3dd9-e34c-4796-b8a1-34c41ece31c1 [ 739.871383] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Successfully created port: 0536866b-6d5c-4ad7-b8b1-afb8ada1d938 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.970615] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.119966] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b54f3b65-dbe0-4eab-80bb-55240e9602cb tempest-MigrationsAdminTest-1026860947 tempest-MigrationsAdminTest-1026860947-project-member] Lock "7b2f3dd9-e34c-4796-b8a1-34c41ece31c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.355s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.389018] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1757c29e-1e81-4d97-b086-f2e27a1550bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.394921] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8bd99c-33aa-480c-b6b6-706729dc84c9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.430563] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f12254f-863b-490e-84a6-19f07071b25b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.438205] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ecec46-49bd-4296-9c5d-009470fbb778 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.451433] env[63197]: DEBUG nova.compute.provider_tree [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.622770] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.801637] env[63197]: DEBUG nova.compute.manager [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Received event network-changed-0536866b-6d5c-4ad7-b8b1-afb8ada1d938 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.801791] env[63197]: DEBUG nova.compute.manager [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Refreshing instance network info cache due to event network-changed-0536866b-6d5c-4ad7-b8b1-afb8ada1d938. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.802438] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] Acquiring lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.802438] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] Acquired lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.802438] env[63197]: DEBUG nova.network.neutron [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Refreshing network info cache for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.955207] env[63197]: DEBUG nova.scheduler.client.report [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.984305] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.986844] env[63197]: ERROR nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. [ 740.986844] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 740.986844] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.986844] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 740.986844] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.986844] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 740.986844] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.986844] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 740.986844] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.986844] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 740.986844] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.986844] env[63197]: ERROR nova.compute.manager raise self.value [ 740.986844] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.986844] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 740.986844] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.986844] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 740.987365] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.987365] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 740.987365] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. [ 740.987365] env[63197]: ERROR nova.compute.manager [ 740.987365] env[63197]: Traceback (most recent call last): [ 740.987365] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 740.987365] env[63197]: listener.cb(fileno) [ 740.987365] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.987365] env[63197]: result = function(*args, **kwargs) [ 740.987365] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.987365] env[63197]: return func(*args, **kwargs) [ 740.987365] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.987365] env[63197]: raise e [ 740.987365] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.987365] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 740.987365] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.987365] env[63197]: created_port_ids = self._update_ports_for_instance( [ 740.987365] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.987365] env[63197]: with excutils.save_and_reraise_exception(): [ 740.987365] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.987365] env[63197]: self.force_reraise() [ 740.987365] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.987365] env[63197]: raise self.value [ 740.987365] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.987365] env[63197]: updated_port = self._update_port( [ 740.987365] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.987365] env[63197]: _ensure_no_port_binding_failure(port) [ 740.987365] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.987365] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 740.988264] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. [ 740.988264] env[63197]: Removing descriptor: 19 [ 741.010407] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.010753] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.010964] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.011361] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.011554] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.011805] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.012151] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.012384] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.012590] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.012788] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.013054] env[63197]: DEBUG nova.virt.hardware [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.014668] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c14a8a-e88a-4746-bf3f-80403f19a7a3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.025525] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89a8503-160f-436c-a88e-0bcf2f943907 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.041252] env[63197]: ERROR nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Traceback (most recent call last): [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] yield resources [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self.driver.spawn(context, instance, image_meta, [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] vm_ref = self.build_virtual_machine(instance, [ 741.041252] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] for vif in network_info: [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] return self._sync_wrapper(fn, *args, **kwargs) [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self.wait() [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self[:] = self._gt.wait() [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] return self._exit_event.wait() [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 741.041663] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] current.throw(*self._exc) [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] result = function(*args, **kwargs) [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] return func(*args, **kwargs) [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] raise e [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] nwinfo = self.network_api.allocate_for_instance( [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] created_port_ids = self._update_ports_for_instance( [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] with excutils.save_and_reraise_exception(): [ 741.042108] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self.force_reraise() [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] raise self.value [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] updated_port = self._update_port( [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] _ensure_no_port_binding_failure(port) [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] raise exception.PortBindingFailed(port_id=port['id']) [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] nova.exception.PortBindingFailed: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. [ 741.042505] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] [ 741.042505] env[63197]: INFO nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Terminating instance [ 741.044239] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Acquiring lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.144936] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.320283] env[63197]: DEBUG nova.network.neutron [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.462038] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.462564] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.465895] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.965s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.467424] env[63197]: INFO nova.compute.claims [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.511312] env[63197]: DEBUG nova.network.neutron [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.974505] env[63197]: DEBUG nova.compute.utils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.982614] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.982834] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 742.014352] env[63197]: DEBUG oslo_concurrency.lockutils [req-4ac37162-2db7-4005-8701-3917e7f031cb req-5b76d8bb-1cdf-482e-abbb-6efc26a9210f service nova] Releasing lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.014710] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Acquired lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.014886] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.027896] env[63197]: DEBUG nova.policy [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '31a519690def4b6b92028150098f7801', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad6294feb57a407d9316ef6bc861262b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.486815] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.542601] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.657293] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Successfully created port: 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.671103] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.904956] env[63197]: DEBUG nova.compute.manager [req-8cc58390-5fb5-430b-b922-a375163cd656 req-3c2083d6-9325-4349-b943-65a3c36f4045 service nova] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Received event network-vif-deleted-0536866b-6d5c-4ad7-b8b1-afb8ada1d938 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.997778] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782cc4be-bc11-46d4-800e-d947c60857c3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.006918] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9a6c04-c21a-49d3-afe9-4389e2c4346c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.046651] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb071c2-ec75-415a-bd4a-064226c2e09b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.055037] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b9c05d-993f-4602-a9a8-26dee19d25d3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.070101] env[63197]: DEBUG nova.compute.provider_tree [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.176121] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Releasing lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.176121] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 743.176121] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.177087] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29f37176-2a1b-4e41-9af5-e74d22153e67 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.186370] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e56fde1-1c8d-4eb8-8667-903c88ac68c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.208824] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 715c53a9-224e-4457-9df9-a2ada9fcae93 could not be found. [ 743.209241] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 743.209550] env[63197]: INFO nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Took 0.03 seconds to destroy the instance on the hypervisor. [ 743.209898] env[63197]: DEBUG oslo.service.loopingcall [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.210355] env[63197]: DEBUG nova.compute.manager [-] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.210554] env[63197]: DEBUG nova.network.neutron [-] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.241690] env[63197]: DEBUG nova.network.neutron [-] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.503385] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.545743] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.547481] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.547724] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.548093] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.548515] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.548930] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.549372] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.549616] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.549853] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.550370] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.550902] env[63197]: DEBUG nova.virt.hardware [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.552867] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5226e334-2b6a-4e56-b176-c0fffb11f66f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.566016] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c946fc5-a728-4015-a107-1315200e5b5c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.583682] env[63197]: DEBUG nova.scheduler.client.report [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.743791] env[63197]: DEBUG nova.network.neutron [-] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.900133] env[63197]: ERROR nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. [ 743.900133] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.900133] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.900133] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.900133] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.900133] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.900133] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.900133] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.900133] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.900133] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 743.900133] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.900133] env[63197]: ERROR nova.compute.manager raise self.value [ 743.900133] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.900133] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.900133] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.900133] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.900869] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.900869] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.900869] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. [ 743.900869] env[63197]: ERROR nova.compute.manager [ 743.900869] env[63197]: Traceback (most recent call last): [ 743.900869] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.900869] env[63197]: listener.cb(fileno) [ 743.900869] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.900869] env[63197]: result = function(*args, **kwargs) [ 743.900869] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.900869] env[63197]: return func(*args, **kwargs) [ 743.900869] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.900869] env[63197]: raise e [ 743.900869] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.900869] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 743.900869] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.900869] env[63197]: created_port_ids = self._update_ports_for_instance( [ 743.900869] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.900869] env[63197]: with excutils.save_and_reraise_exception(): [ 743.900869] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.900869] env[63197]: self.force_reraise() [ 743.900869] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.900869] env[63197]: raise self.value [ 743.900869] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.900869] env[63197]: updated_port = self._update_port( [ 743.900869] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.900869] env[63197]: _ensure_no_port_binding_failure(port) [ 743.900869] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.900869] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.901676] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. [ 743.901676] env[63197]: Removing descriptor: 19 [ 743.901676] env[63197]: ERROR nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Traceback (most recent call last): [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] yield resources [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self.driver.spawn(context, instance, image_meta, [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.901676] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] vm_ref = self.build_virtual_machine(instance, [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] for vif in network_info: [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return self._sync_wrapper(fn, *args, **kwargs) [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self.wait() [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self[:] = self._gt.wait() [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return self._exit_event.wait() [ 743.902150] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] result = hub.switch() [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return self.greenlet.switch() [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] result = function(*args, **kwargs) [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return func(*args, **kwargs) [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] raise e [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] nwinfo = self.network_api.allocate_for_instance( [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.902500] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] created_port_ids = self._update_ports_for_instance( [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] with excutils.save_and_reraise_exception(): [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self.force_reraise() [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] raise self.value [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] updated_port = self._update_port( [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] _ensure_no_port_binding_failure(port) [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.902866] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] raise exception.PortBindingFailed(port_id=port['id']) [ 743.903209] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] nova.exception.PortBindingFailed: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. [ 743.903209] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] [ 743.903209] env[63197]: INFO nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Terminating instance [ 743.904512] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.904512] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquired lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.904512] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.087766] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.088312] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.092622] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.591s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.249031] env[63197]: INFO nova.compute.manager [-] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Took 1.04 seconds to deallocate network for instance. [ 744.251105] env[63197]: DEBUG nova.compute.claims [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 744.251302] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.425326] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.574021] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.598184] env[63197]: DEBUG nova.compute.utils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.604441] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.604441] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 744.678786] env[63197]: DEBUG nova.policy [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0ff59d58aa114bc991ac25218c61cab3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'afe23b33db354d6caca0b167256e560a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.976532] env[63197]: DEBUG nova.compute.manager [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Received event network-changed-5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.976872] env[63197]: DEBUG nova.compute.manager [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Refreshing instance network info cache due to event network-changed-5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.977182] env[63197]: DEBUG oslo_concurrency.lockutils [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] Acquiring lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.055359] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82fa53c1-8d1e-43bb-bdb7-9019987f83ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.063345] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87011e8-1794-4bec-938d-a7319150f0e7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.103117] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Releasing lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.103117] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 745.103117] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 745.103117] env[63197]: DEBUG oslo_concurrency.lockutils [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] Acquired lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.103117] env[63197]: DEBUG nova.network.neutron [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Refreshing network info cache for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 745.104740] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-658cdf3c-8670-4143-b0f7-42c332e021c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.113272] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f302e8-ba4a-41cc-9ac5-b49f817207da {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.113568] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.124111] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d545c06f-d556-4293-aada-12b02b35b636 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.131459] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beac8ac3-39b9-40d4-a697-7d1269553360 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.152299] env[63197]: DEBUG nova.compute.provider_tree [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.162733] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 86a728c9-27d3-4c98-a4d6-6b0736f24648 could not be found. [ 745.162953] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.163145] env[63197]: INFO nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Took 0.06 seconds to destroy the instance on the hypervisor. [ 745.163378] env[63197]: DEBUG oslo.service.loopingcall [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.164308] env[63197]: DEBUG nova.scheduler.client.report [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.167947] env[63197]: DEBUG nova.compute.manager [-] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.168068] env[63197]: DEBUG nova.network.neutron [-] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.183649] env[63197]: DEBUG nova.network.neutron [-] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.253032] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Successfully created port: 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.637067] env[63197]: DEBUG nova.network.neutron [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.672145] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.580s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.672251] env[63197]: ERROR nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Traceback (most recent call last): [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self.driver.spawn(context, instance, image_meta, [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] vm_ref = self.build_virtual_machine(instance, [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.672251] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] for vif in network_info: [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return self._sync_wrapper(fn, *args, **kwargs) [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self.wait() [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self[:] = self._gt.wait() [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return self._exit_event.wait() [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] result = hub.switch() [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.672600] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return self.greenlet.switch() [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] result = function(*args, **kwargs) [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] return func(*args, **kwargs) [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] raise e [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] nwinfo = self.network_api.allocate_for_instance( [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] created_port_ids = self._update_ports_for_instance( [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] with excutils.save_and_reraise_exception(): [ 745.672974] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] self.force_reraise() [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] raise self.value [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] updated_port = self._update_port( [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] _ensure_no_port_binding_failure(port) [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] raise exception.PortBindingFailed(port_id=port['id']) [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] nova.exception.PortBindingFailed: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. [ 745.673430] env[63197]: ERROR nova.compute.manager [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] [ 745.673744] env[63197]: DEBUG nova.compute.utils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.674337] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.595s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.675717] env[63197]: INFO nova.compute.claims [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.678503] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Build of instance 1e650df0-b4a3-45dd-8b14-b41d4c77b625 was re-scheduled: Binding failed for port 02e94589-6b93-48ed-913a-58f23556b583, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.678836] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.679126] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Acquiring lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.679277] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Acquired lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.679434] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.686544] env[63197]: DEBUG nova.network.neutron [-] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.768522] env[63197]: DEBUG nova.network.neutron [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.133337] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.163896] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.164149] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.164298] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.164475] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.164613] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.164764] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.164967] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.165209] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.165394] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.165552] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.165717] env[63197]: DEBUG nova.virt.hardware [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.166603] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f84906-0eb6-498e-80ea-3fd92f429921 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.175268] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0921b17-267a-4cc1-bef2-71552a42794e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.199930] env[63197]: INFO nova.compute.manager [-] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Took 1.03 seconds to deallocate network for instance. [ 746.203103] env[63197]: DEBUG nova.compute.claims [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 746.203286] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.223311] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.273642] env[63197]: DEBUG oslo_concurrency.lockutils [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] Releasing lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.274723] env[63197]: DEBUG nova.compute.manager [req-a54e7bee-7387-436f-828d-fd3b7bf11a6a req-669a0e14-c422-4cbc-bd88-6ac403fa8bfc service nova] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Received event network-vif-deleted-5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.320850] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.507259] env[63197]: ERROR nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. [ 746.507259] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 746.507259] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.507259] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 746.507259] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.507259] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 746.507259] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.507259] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 746.507259] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.507259] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 746.507259] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.507259] env[63197]: ERROR nova.compute.manager raise self.value [ 746.507259] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.507259] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 746.507259] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.507259] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 746.507701] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.507701] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 746.507701] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. [ 746.507701] env[63197]: ERROR nova.compute.manager [ 746.507701] env[63197]: Traceback (most recent call last): [ 746.507701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 746.507701] env[63197]: listener.cb(fileno) [ 746.507701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.507701] env[63197]: result = function(*args, **kwargs) [ 746.507701] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.507701] env[63197]: return func(*args, **kwargs) [ 746.507701] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.507701] env[63197]: raise e [ 746.507701] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.507701] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 746.507701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.507701] env[63197]: created_port_ids = self._update_ports_for_instance( [ 746.507701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.507701] env[63197]: with excutils.save_and_reraise_exception(): [ 746.507701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.507701] env[63197]: self.force_reraise() [ 746.507701] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.507701] env[63197]: raise self.value [ 746.507701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.507701] env[63197]: updated_port = self._update_port( [ 746.507701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.507701] env[63197]: _ensure_no_port_binding_failure(port) [ 746.507701] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.507701] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 746.508490] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. [ 746.508490] env[63197]: Removing descriptor: 19 [ 746.508490] env[63197]: ERROR nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Traceback (most recent call last): [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] yield resources [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self.driver.spawn(context, instance, image_meta, [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.508490] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] vm_ref = self.build_virtual_machine(instance, [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] for vif in network_info: [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return self._sync_wrapper(fn, *args, **kwargs) [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self.wait() [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self[:] = self._gt.wait() [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return self._exit_event.wait() [ 746.508910] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] result = hub.switch() [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return self.greenlet.switch() [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] result = function(*args, **kwargs) [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return func(*args, **kwargs) [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] raise e [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] nwinfo = self.network_api.allocate_for_instance( [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.509262] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] created_port_ids = self._update_ports_for_instance( [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] with excutils.save_and_reraise_exception(): [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self.force_reraise() [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] raise self.value [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] updated_port = self._update_port( [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] _ensure_no_port_binding_failure(port) [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.509602] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] raise exception.PortBindingFailed(port_id=port['id']) [ 746.509920] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] nova.exception.PortBindingFailed: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. [ 746.509920] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] [ 746.509920] env[63197]: INFO nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Terminating instance [ 746.513030] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Acquiring lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.513030] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Acquired lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.513192] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.652181] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.652181] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.732319] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.732568] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.822240] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Releasing lock "refresh_cache-1e650df0-b4a3-45dd-8b14-b41d4c77b625" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.822478] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.822641] env[63197]: DEBUG nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.823101] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.844399] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.002872] env[63197]: DEBUG nova.compute.manager [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Received event network-changed-8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.002872] env[63197]: DEBUG nova.compute.manager [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Refreshing instance network info cache due to event network-changed-8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 747.003623] env[63197]: DEBUG oslo_concurrency.lockutils [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] Acquiring lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.032429] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.074309] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c61c5a-fe36-4ac0-af84-ac7f7433209f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.084321] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daca31de-b3ee-49a3-ae5d-d444168cd5d3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.115174] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f69d9c7-d974-4f22-9000-fe21b81e2f08 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.118349] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.125140] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c4ef42-bf82-49db-bcf1-5f9817ae3f6c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.139421] env[63197]: DEBUG nova.compute.provider_tree [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.347483] env[63197]: DEBUG nova.network.neutron [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.622008] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Releasing lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.623219] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.623219] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.623219] env[63197]: DEBUG oslo_concurrency.lockutils [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] Acquired lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.623575] env[63197]: DEBUG nova.network.neutron [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Refreshing network info cache for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.626044] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-437b0dde-e629-4557-adc0-bfd2ca0bb569 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.637130] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9027ff8d-ced8-4f74-bdde-f540feb25831 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.654604] env[63197]: DEBUG nova.scheduler.client.report [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.671115] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 218ad539-e2d0-4d4b-9857-ec9d2af0be03 could not be found. [ 747.671115] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 747.671115] env[63197]: INFO nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Took 0.05 seconds to destroy the instance on the hypervisor. [ 747.671115] env[63197]: DEBUG oslo.service.loopingcall [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.671769] env[63197]: DEBUG nova.compute.manager [-] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 747.672029] env[63197]: DEBUG nova.network.neutron [-] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.741096] env[63197]: DEBUG nova.network.neutron [-] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.852050] env[63197]: INFO nova.compute.manager [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] [instance: 1e650df0-b4a3-45dd-8b14-b41d4c77b625] Took 1.03 seconds to deallocate network for instance. [ 748.150059] env[63197]: DEBUG nova.network.neutron [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.160596] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.163017] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.164510] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.331s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.164969] env[63197]: DEBUG nova.objects.instance [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lazy-loading 'resources' on Instance uuid 09dc9549-ab9a-4f36-9671-ab1b5595649d {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.245671] env[63197]: DEBUG nova.network.neutron [-] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.247526] env[63197]: DEBUG nova.network.neutron [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.672333] env[63197]: DEBUG nova.compute.utils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.674021] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.674225] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.723828] env[63197]: DEBUG nova.policy [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11cafcbba38444e4a1878967db8c7f7f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ba27de188e814ad5a39404212440e3cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.750592] env[63197]: DEBUG oslo_concurrency.lockutils [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] Releasing lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.750858] env[63197]: DEBUG nova.compute.manager [req-9e6b2df0-2077-453c-9268-500bc2c94663 req-0b51997d-b73e-4776-b5bd-dd3cc93483d1 service nova] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Received event network-vif-deleted-8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.751497] env[63197]: INFO nova.compute.manager [-] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Took 1.08 seconds to deallocate network for instance. [ 748.753757] env[63197]: DEBUG nova.compute.claims [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 748.753928] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.879887] env[63197]: INFO nova.scheduler.client.report [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Deleted allocations for instance 1e650df0-b4a3-45dd-8b14-b41d4c77b625 [ 749.037222] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Successfully created port: 884ecdf1-7984-42fc-a266-49330148a117 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.063345] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d799cb64-5b21-4a65-9215-4e5dc3c90cbc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.070686] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9858425e-595a-4212-9d5b-f25e789e6201 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.099477] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e76e12-3b22-4a4f-86e1-323939c83627 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.106275] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b253781-9e2b-47d4-94de-dd811d2af382 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.120074] env[63197]: DEBUG nova.compute.provider_tree [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.177869] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.393365] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2c183690-030d-407a-a042-873425870aba tempest-ServerExternalEventsTest-174949277 tempest-ServerExternalEventsTest-174949277-project-member] Lock "1e650df0-b4a3-45dd-8b14-b41d4c77b625" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.339s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.623722] env[63197]: DEBUG nova.scheduler.client.report [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.831592] env[63197]: DEBUG nova.compute.manager [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Received event network-changed-884ecdf1-7984-42fc-a266-49330148a117 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.831848] env[63197]: DEBUG nova.compute.manager [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Refreshing instance network info cache due to event network-changed-884ecdf1-7984-42fc-a266-49330148a117. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 749.832092] env[63197]: DEBUG oslo_concurrency.lockutils [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] Acquiring lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.832241] env[63197]: DEBUG oslo_concurrency.lockutils [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] Acquired lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.832402] env[63197]: DEBUG nova.network.neutron [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Refreshing network info cache for port 884ecdf1-7984-42fc-a266-49330148a117 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.896515] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 750.071560] env[63197]: ERROR nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. [ 750.071560] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.071560] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.071560] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.071560] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.071560] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.071560] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.071560] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.071560] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.071560] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 750.071560] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.071560] env[63197]: ERROR nova.compute.manager raise self.value [ 750.071560] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.071560] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.071560] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.071560] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.072221] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.072221] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.072221] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. [ 750.072221] env[63197]: ERROR nova.compute.manager [ 750.072221] env[63197]: Traceback (most recent call last): [ 750.072221] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.072221] env[63197]: listener.cb(fileno) [ 750.072221] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.072221] env[63197]: result = function(*args, **kwargs) [ 750.072221] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.072221] env[63197]: return func(*args, **kwargs) [ 750.072221] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.072221] env[63197]: raise e [ 750.072221] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.072221] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 750.072221] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.072221] env[63197]: created_port_ids = self._update_ports_for_instance( [ 750.072221] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.072221] env[63197]: with excutils.save_and_reraise_exception(): [ 750.072221] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.072221] env[63197]: self.force_reraise() [ 750.072221] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.072221] env[63197]: raise self.value [ 750.072221] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.072221] env[63197]: updated_port = self._update_port( [ 750.072221] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.072221] env[63197]: _ensure_no_port_binding_failure(port) [ 750.072221] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.072221] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.073012] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. [ 750.073012] env[63197]: Removing descriptor: 19 [ 750.132535] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.967s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.135043] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.136s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.159873] env[63197]: INFO nova.scheduler.client.report [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Deleted allocations for instance 09dc9549-ab9a-4f36-9671-ab1b5595649d [ 750.188978] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.213018] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.213289] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.213445] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.213623] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.213768] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.213914] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.214128] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.214284] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.214446] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.214605] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.214777] env[63197]: DEBUG nova.virt.hardware [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.215645] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300d819e-b192-4e37-982f-5581004dc67f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.223750] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35e532d-8411-4134-806e-b094199920a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.237594] env[63197]: ERROR nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Traceback (most recent call last): [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] yield resources [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self.driver.spawn(context, instance, image_meta, [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] vm_ref = self.build_virtual_machine(instance, [ 750.237594] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] for vif in network_info: [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] return self._sync_wrapper(fn, *args, **kwargs) [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self.wait() [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self[:] = self._gt.wait() [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] return self._exit_event.wait() [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.238116] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] current.throw(*self._exc) [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] result = function(*args, **kwargs) [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] return func(*args, **kwargs) [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] raise e [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] nwinfo = self.network_api.allocate_for_instance( [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] created_port_ids = self._update_ports_for_instance( [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] with excutils.save_and_reraise_exception(): [ 750.238517] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self.force_reraise() [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] raise self.value [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] updated_port = self._update_port( [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] _ensure_no_port_binding_failure(port) [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] raise exception.PortBindingFailed(port_id=port['id']) [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] nova.exception.PortBindingFailed: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. [ 750.238905] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] [ 750.238905] env[63197]: INFO nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Terminating instance [ 750.239667] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Acquiring lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.349642] env[63197]: DEBUG nova.network.neutron [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 750.417894] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.444181] env[63197]: DEBUG nova.network.neutron [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.667010] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6619841-3de6-4aa4-864a-c95c5cf3c0b0 tempest-ServersAaction247Test-1244557579 tempest-ServersAaction247Test-1244557579-project-member] Lock "09dc9549-ab9a-4f36-9671-ab1b5595649d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.523s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.946600] env[63197]: DEBUG oslo_concurrency.lockutils [req-66d22f42-6407-40ee-994b-99743501046b req-591f53ac-1da4-4463-b0c8-09a40671e7d6 service nova] Releasing lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.946883] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Acquired lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.947091] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.997911] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bd52c7-10b6-498e-801f-800d1c71b29f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.005494] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4392d9-0798-4d8c-8c28-6126fc07a14c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.039213] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69c8f3a-5aa6-4022-b258-dc8369866278 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.048130] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8050a7eb-f1cb-4bf1-af51-12783d3f1edd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.059736] env[63197]: DEBUG nova.compute.provider_tree [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.484331] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.562877] env[63197]: DEBUG nova.scheduler.client.report [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.647668] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.885388] env[63197]: DEBUG nova.compute.manager [req-27f15b21-720a-4367-b44e-e18dfb747bd4 req-aeec3206-3c3f-4213-b633-62c897aab7d6 service nova] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Received event network-vif-deleted-884ecdf1-7984-42fc-a266-49330148a117 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.069750] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.070442] env[63197]: ERROR nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Traceback (most recent call last): [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self.driver.spawn(context, instance, image_meta, [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] vm_ref = self.build_virtual_machine(instance, [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.070442] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] for vif in network_info: [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] return self._sync_wrapper(fn, *args, **kwargs) [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self.wait() [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self[:] = self._gt.wait() [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] return self._exit_event.wait() [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] current.throw(*self._exc) [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.070831] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] result = function(*args, **kwargs) [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] return func(*args, **kwargs) [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] raise e [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] nwinfo = self.network_api.allocate_for_instance( [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] created_port_ids = self._update_ports_for_instance( [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] with excutils.save_and_reraise_exception(): [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] self.force_reraise() [ 752.071425] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] raise self.value [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] updated_port = self._update_port( [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] _ensure_no_port_binding_failure(port) [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] raise exception.PortBindingFailed(port_id=port['id']) [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] nova.exception.PortBindingFailed: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. [ 752.071928] env[63197]: ERROR nova.compute.manager [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] [ 752.071928] env[63197]: DEBUG nova.compute.utils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.072617] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.349s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.075504] env[63197]: INFO nova.compute.claims [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.078161] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Build of instance 0f566f8c-c92a-4885-94a6-08043857a1c2 was re-scheduled: Binding failed for port fc86d49c-c9ff-4aef-badc-b94cd8d432bf, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 752.078596] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 752.078920] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.079009] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquired lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.079171] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 752.150401] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Releasing lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.150906] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.151137] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.151663] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af301387-3d6a-439b-89e7-d692562e1e3f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.161107] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51cb37a-efa4-4e7e-a6a3-1fcd963c07c2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.181499] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d9f1dcc1-77fd-4c8c-b930-eb421691dc94 could not be found. [ 752.181708] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.181913] env[63197]: INFO nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.182212] env[63197]: DEBUG oslo.service.loopingcall [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.182437] env[63197]: DEBUG nova.compute.manager [-] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.182532] env[63197]: DEBUG nova.network.neutron [-] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.196157] env[63197]: DEBUG nova.network.neutron [-] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.614695] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.624705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.624935] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.698789] env[63197]: DEBUG nova.network.neutron [-] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.734152] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.202958] env[63197]: INFO nova.compute.manager [-] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Took 1.02 seconds to deallocate network for instance. [ 753.208205] env[63197]: DEBUG nova.compute.claims [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.208417] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.238047] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Releasing lock "refresh_cache-0f566f8c-c92a-4885-94a6-08043857a1c2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.238290] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 753.238465] env[63197]: DEBUG nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.238642] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 753.258288] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.469427] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96a6cba-9110-417d-8980-a7a3da407241 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.476983] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc17f1a-1d1e-4fc2-ba6d-74257130206e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.513297] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d31db53-8e73-4b78-bc17-eb7de1ab6f6b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.521520] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8a2758-98f2-433a-bbed-16abd8be7a4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.536871] env[63197]: DEBUG nova.compute.provider_tree [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.764021] env[63197]: DEBUG nova.network.neutron [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.040654] env[63197]: DEBUG nova.scheduler.client.report [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.265169] env[63197]: INFO nova.compute.manager [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 0f566f8c-c92a-4885-94a6-08043857a1c2] Took 1.03 seconds to deallocate network for instance. [ 754.547626] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.548171] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.550866] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.830s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.057889] env[63197]: DEBUG nova.compute.utils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.065491] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.065491] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 755.116694] env[63197]: DEBUG nova.policy [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1337093c94e4267879f93efbfb46303', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '065bd4a13f694fd0aa6c1d482991638c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.298511] env[63197]: INFO nova.scheduler.client.report [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Deleted allocations for instance 0f566f8c-c92a-4885-94a6-08043857a1c2 [ 755.473151] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a4fe6c-ee87-420a-a489-3d513d5ae9fd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.484508] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b26d52-0aea-41e6-8fea-e022d85156db {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.521084] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7f3a90-e9a9-4996-af5e-1ac4607dfaf5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.528849] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1382bcdc-951d-4e46-903c-cb022bf98a69 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.541954] env[63197]: DEBUG nova.compute.provider_tree [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.560147] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Successfully created port: 5afd3f1b-0841-41b6-b92e-10497d48f1e6 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.563343] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.810336] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1898f2ae-9a7a-452f-9efe-0a6a63cbc45c tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "0f566f8c-c92a-4885-94a6-08043857a1c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.873s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.046406] env[63197]: DEBUG nova.scheduler.client.report [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.313754] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 756.552279] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.553013] env[63197]: ERROR nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Traceback (most recent call last): [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self.driver.spawn(context, instance, image_meta, [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] vm_ref = self.build_virtual_machine(instance, [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.553013] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] for vif in network_info: [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return self._sync_wrapper(fn, *args, **kwargs) [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self.wait() [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self[:] = self._gt.wait() [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return self._exit_event.wait() [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] result = hub.switch() [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 756.553464] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return self.greenlet.switch() [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] result = function(*args, **kwargs) [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] return func(*args, **kwargs) [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] raise e [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] nwinfo = self.network_api.allocate_for_instance( [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] created_port_ids = self._update_ports_for_instance( [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] with excutils.save_and_reraise_exception(): [ 756.553878] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] self.force_reraise() [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] raise self.value [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] updated_port = self._update_port( [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] _ensure_no_port_binding_failure(port) [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] raise exception.PortBindingFailed(port_id=port['id']) [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] nova.exception.PortBindingFailed: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. [ 756.554316] env[63197]: ERROR nova.compute.manager [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] [ 756.554713] env[63197]: DEBUG nova.compute.utils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 756.555473] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Build of instance a0638d5a-6d33-4eb6-a9b6-ad6620219259 was re-scheduled: Binding failed for port 35397525-7932-47df-a882-adda1487c79b, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 756.555883] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 756.556153] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.556303] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.556462] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.558903] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.329s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.560497] env[63197]: INFO nova.compute.claims [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.575702] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.605445] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.605704] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.605836] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.606048] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.606196] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.606342] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.606549] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.606709] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.606878] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.607403] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.607632] env[63197]: DEBUG nova.virt.hardware [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.611089] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfdda12-29f4-44d0-9be6-adb87be4c169 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.619550] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc95646-1380-4972-a930-3d2ce16cf2cd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.727933] env[63197]: DEBUG nova.compute.manager [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Received event network-changed-5afd3f1b-0841-41b6-b92e-10497d48f1e6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.728118] env[63197]: DEBUG nova.compute.manager [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Refreshing instance network info cache due to event network-changed-5afd3f1b-0841-41b6-b92e-10497d48f1e6. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 756.728343] env[63197]: DEBUG oslo_concurrency.lockutils [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] Acquiring lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.728482] env[63197]: DEBUG oslo_concurrency.lockutils [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] Acquired lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.728664] env[63197]: DEBUG nova.network.neutron [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Refreshing network info cache for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 756.842740] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.089156] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.097503] env[63197]: ERROR nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. [ 757.097503] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.097503] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.097503] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.097503] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.097503] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.097503] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.097503] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.097503] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.097503] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 757.097503] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.097503] env[63197]: ERROR nova.compute.manager raise self.value [ 757.097503] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.097503] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.097503] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.097503] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.097998] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.097998] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.097998] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. [ 757.097998] env[63197]: ERROR nova.compute.manager [ 757.097998] env[63197]: Traceback (most recent call last): [ 757.097998] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.097998] env[63197]: listener.cb(fileno) [ 757.097998] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.097998] env[63197]: result = function(*args, **kwargs) [ 757.097998] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.097998] env[63197]: return func(*args, **kwargs) [ 757.097998] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.097998] env[63197]: raise e [ 757.097998] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.097998] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 757.097998] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.097998] env[63197]: created_port_ids = self._update_ports_for_instance( [ 757.097998] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.097998] env[63197]: with excutils.save_and_reraise_exception(): [ 757.097998] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.097998] env[63197]: self.force_reraise() [ 757.097998] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.097998] env[63197]: raise self.value [ 757.097998] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.097998] env[63197]: updated_port = self._update_port( [ 757.097998] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.097998] env[63197]: _ensure_no_port_binding_failure(port) [ 757.097998] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.097998] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.098786] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. [ 757.098786] env[63197]: Removing descriptor: 19 [ 757.098786] env[63197]: ERROR nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Traceback (most recent call last): [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] yield resources [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self.driver.spawn(context, instance, image_meta, [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.098786] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] vm_ref = self.build_virtual_machine(instance, [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] for vif in network_info: [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return self._sync_wrapper(fn, *args, **kwargs) [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self.wait() [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self[:] = self._gt.wait() [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return self._exit_event.wait() [ 757.099150] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] result = hub.switch() [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return self.greenlet.switch() [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] result = function(*args, **kwargs) [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return func(*args, **kwargs) [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] raise e [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] nwinfo = self.network_api.allocate_for_instance( [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.099500] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] created_port_ids = self._update_ports_for_instance( [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] with excutils.save_and_reraise_exception(): [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self.force_reraise() [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] raise self.value [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] updated_port = self._update_port( [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] _ensure_no_port_binding_failure(port) [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.099857] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] raise exception.PortBindingFailed(port_id=port['id']) [ 757.100261] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] nova.exception.PortBindingFailed: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. [ 757.100261] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] [ 757.100261] env[63197]: INFO nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Terminating instance [ 757.102237] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Acquiring lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.191793] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.258175] env[63197]: DEBUG nova.network.neutron [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.342448] env[63197]: DEBUG nova.network.neutron [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.695530] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-a0638d5a-6d33-4eb6-a9b6-ad6620219259" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.695792] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 757.696065] env[63197]: DEBUG nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.696394] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 757.717954] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.845429] env[63197]: DEBUG oslo_concurrency.lockutils [req-9a041530-8945-42b0-9ca8-0c8da4070384 req-ceab99d1-5cf0-4381-bcae-952a054dac72 service nova] Releasing lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.847990] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Acquired lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.848286] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.917467] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8659b6-c21d-4225-8ef9-616d11ef5472 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.925640] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dcc74f-f77e-49dc-8897-cae7eb9e196e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.958240] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74c6b3d-3112-4784-b458-30521512cf01 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.966093] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d067b79-c74b-4f99-9dc3-c963f76761c3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.980551] env[63197]: DEBUG nova.compute.provider_tree [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.219575] env[63197]: DEBUG nova.network.neutron [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.373320] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.462054] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.484111] env[63197]: DEBUG nova.scheduler.client.report [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.722516] env[63197]: INFO nova.compute.manager [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: a0638d5a-6d33-4eb6-a9b6-ad6620219259] Took 1.03 seconds to deallocate network for instance. [ 758.754824] env[63197]: DEBUG nova.compute.manager [req-3daa0131-1b61-4d1b-a8ff-a2c51f006fa7 req-d802fe73-f3db-4ad8-8512-9a71e0439616 service nova] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Received event network-vif-deleted-5afd3f1b-0841-41b6-b92e-10497d48f1e6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.964689] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Releasing lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.965158] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 758.965356] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.965660] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f52ce05-7ce6-4944-9934-2a8360df5f2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.974700] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff412eb-fd0a-4ba1-88e6-77cdaa6c9e98 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.995324] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.995665] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.998548] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1ba5772-9e73-4a51-ae62-d9c17a64ff0d could not be found. [ 758.999134] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.999134] env[63197]: INFO nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 758.999303] env[63197]: DEBUG oslo.service.loopingcall [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.999757] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.855s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.001156] env[63197]: INFO nova.compute.claims [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.003667] env[63197]: DEBUG nova.compute.manager [-] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.003920] env[63197]: DEBUG nova.network.neutron [-] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.019493] env[63197]: DEBUG nova.network.neutron [-] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.501623] env[63197]: DEBUG nova.compute.utils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.504674] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.505577] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.522034] env[63197]: DEBUG nova.network.neutron [-] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.549436] env[63197]: DEBUG nova.policy [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0607a527a0b4482b8cd6beba3818d0d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9752164b6c745ad87f27e3adf353e3a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.762925] env[63197]: INFO nova.scheduler.client.report [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleted allocations for instance a0638d5a-6d33-4eb6-a9b6-ad6620219259 [ 759.880564] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Successfully created port: b936185c-ff52-4611-8304-2d440945a78f {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.008620] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.024432] env[63197]: INFO nova.compute.manager [-] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Took 1.02 seconds to deallocate network for instance. [ 760.026698] env[63197]: DEBUG nova.compute.claims [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 760.026879] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.275551] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e48ec8ff-dbe4-42bb-9c63-0ae4415e18d3 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "a0638d5a-6d33-4eb6-a9b6-ad6620219259" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.553s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.313649] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2881be3a-373e-4edd-9182-9df71f769a70 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.320994] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03cc2fe2-126e-45df-8e35-165b3dd0db4b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.356863] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c3f8a0-b21d-4880-86e2-5d8abe068966 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.364525] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7f1d6f-d8d5-4be1-bbf9-77b5f3964fc7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.377968] env[63197]: DEBUG nova.compute.provider_tree [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.516056] env[63197]: INFO nova.virt.block_device [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Booting with volume c007b7d2-cbb2-4521-b7a3-a76059f23352 at /dev/sda [ 760.574571] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b5ee8eba-c9be-4c75-95ec-dbdb38ee8c6f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.583870] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a5e5e2-7178-4413-9588-d1f909f4fc19 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.604218] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3de7788c-b79d-4425-9a89-d8ef5b9e1e94 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.612137] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ac6ab7-29eb-4f53-bc20-aa088ecb636e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.632521] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792cfa83-d99f-46ef-a87a-6ca1911c3ab5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.639141] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1d684c-4ac5-43e1-9bd1-0533d4a59d00 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.652377] env[63197]: DEBUG nova.virt.block_device [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Updating existing volume attachment record: 10e1c1f4-c4f7-4e29-a4a3-5e60e43ed17d {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 760.778346] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.881206] env[63197]: DEBUG nova.scheduler.client.report [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.904895] env[63197]: DEBUG nova.compute.manager [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Received event network-changed-b936185c-ff52-4611-8304-2d440945a78f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.905108] env[63197]: DEBUG nova.compute.manager [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Refreshing instance network info cache due to event network-changed-b936185c-ff52-4611-8304-2d440945a78f. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 760.905325] env[63197]: DEBUG oslo_concurrency.lockutils [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] Acquiring lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.905465] env[63197]: DEBUG oslo_concurrency.lockutils [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] Acquired lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.905637] env[63197]: DEBUG nova.network.neutron [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Refreshing network info cache for port b936185c-ff52-4611-8304-2d440945a78f {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 761.071106] env[63197]: ERROR nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. [ 761.071106] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.071106] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.071106] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.071106] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.071106] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.071106] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.071106] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.071106] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.071106] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 761.071106] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.071106] env[63197]: ERROR nova.compute.manager raise self.value [ 761.071106] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.071106] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.071106] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.071106] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.071758] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.071758] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.071758] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. [ 761.071758] env[63197]: ERROR nova.compute.manager [ 761.071758] env[63197]: Traceback (most recent call last): [ 761.071758] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.071758] env[63197]: listener.cb(fileno) [ 761.071758] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.071758] env[63197]: result = function(*args, **kwargs) [ 761.071758] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.071758] env[63197]: return func(*args, **kwargs) [ 761.071758] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.071758] env[63197]: raise e [ 761.071758] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.071758] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 761.071758] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.071758] env[63197]: created_port_ids = self._update_ports_for_instance( [ 761.071758] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.071758] env[63197]: with excutils.save_and_reraise_exception(): [ 761.071758] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.071758] env[63197]: self.force_reraise() [ 761.071758] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.071758] env[63197]: raise self.value [ 761.071758] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.071758] env[63197]: updated_port = self._update_port( [ 761.071758] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.071758] env[63197]: _ensure_no_port_binding_failure(port) [ 761.071758] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.071758] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.072692] env[63197]: nova.exception.PortBindingFailed: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. [ 761.072692] env[63197]: Removing descriptor: 18 [ 761.305308] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.386060] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.386s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.386587] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.389121] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.138s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.423697] env[63197]: DEBUG nova.network.neutron [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.550185] env[63197]: DEBUG nova.network.neutron [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.893424] env[63197]: DEBUG nova.compute.utils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 761.896438] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 761.896438] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 761.966326] env[63197]: DEBUG nova.policy [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccf7a4049c9d4341a59b7a050a1d5f09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d78a0bec6e64f7dacabbf83f0b1b13f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.052663] env[63197]: DEBUG oslo_concurrency.lockutils [req-95b0ffb2-1505-4047-b096-c047068c1435 req-f57e59f6-48e6-4876-8aa1-672fae828303 service nova] Releasing lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.288585] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5791da6b-1ab0-4651-b49c-9500ceed2d19 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.296467] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c7ae36a-f4ae-4e3c-8e36-146b44b0a323 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.326078] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbd60ff-6135-43a6-80e8-eb616519adff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.333969] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fa581b-b8de-4c5c-bc17-93a3bdb4e829 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.349400] env[63197]: DEBUG nova.compute.provider_tree [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.402516] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.430928] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Successfully created port: e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.775118] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 762.775700] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.775908] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.776076] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.776280] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.776398] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.776537] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.776737] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.776893] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.777066] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.777227] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.777393] env[63197]: DEBUG nova.virt.hardware [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.778256] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e8e412-8693-4a74-ad6b-6c8a104369a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.786732] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4e2e02-d600-4dd6-a5c7-a7d67501191a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.800644] env[63197]: ERROR nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Traceback (most recent call last): [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] yield resources [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self.driver.spawn(context, instance, image_meta, [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] vm_ref = self.build_virtual_machine(instance, [ 762.800644] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] for vif in network_info: [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] return self._sync_wrapper(fn, *args, **kwargs) [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self.wait() [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self[:] = self._gt.wait() [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] return self._exit_event.wait() [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 762.801107] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] current.throw(*self._exc) [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] result = function(*args, **kwargs) [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] return func(*args, **kwargs) [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] raise e [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] nwinfo = self.network_api.allocate_for_instance( [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] created_port_ids = self._update_ports_for_instance( [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] with excutils.save_and_reraise_exception(): [ 762.801530] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self.force_reraise() [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] raise self.value [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] updated_port = self._update_port( [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] _ensure_no_port_binding_failure(port) [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] raise exception.PortBindingFailed(port_id=port['id']) [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] nova.exception.PortBindingFailed: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. [ 762.801968] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] [ 762.801968] env[63197]: INFO nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Terminating instance [ 762.803219] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Acquiring lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.803376] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Acquired lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.803540] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.852744] env[63197]: DEBUG nova.scheduler.client.report [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.051433] env[63197]: DEBUG nova.compute.manager [req-37647c97-9cd1-46f9-9cf2-8226da44a2ea req-2fbf490c-6049-490d-a387-ef71183d6477 service nova] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Received event network-vif-deleted-b936185c-ff52-4611-8304-2d440945a78f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 763.321651] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.359254] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.359254] env[63197]: ERROR nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. [ 763.359254] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Traceback (most recent call last): [ 763.359254] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.359254] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self.driver.spawn(context, instance, image_meta, [ 763.359254] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 763.359254] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.359254] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.359254] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] vm_ref = self.build_virtual_machine(instance, [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] for vif in network_info: [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] return self._sync_wrapper(fn, *args, **kwargs) [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self.wait() [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self[:] = self._gt.wait() [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] return self._exit_event.wait() [ 763.359677] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] current.throw(*self._exc) [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] result = function(*args, **kwargs) [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] return func(*args, **kwargs) [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] raise e [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] nwinfo = self.network_api.allocate_for_instance( [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] created_port_ids = self._update_ports_for_instance( [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.360053] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] with excutils.save_and_reraise_exception(): [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] self.force_reraise() [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] raise self.value [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] updated_port = self._update_port( [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] _ensure_no_port_binding_failure(port) [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] raise exception.PortBindingFailed(port_id=port['id']) [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] nova.exception.PortBindingFailed: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. [ 763.360402] env[63197]: ERROR nova.compute.manager [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] [ 763.360773] env[63197]: DEBUG nova.compute.utils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 763.360773] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.156s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.364049] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Build of instance 715c53a9-224e-4457-9df9-a2ada9fcae93 was re-scheduled: Binding failed for port 0536866b-6d5c-4ad7-b8b1-afb8ada1d938, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 763.364295] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 763.364511] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Acquiring lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.364654] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Acquired lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.364838] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.399044] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.413783] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.438244] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.438244] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.438244] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.439730] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.439730] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.439730] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.439730] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.439730] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.439919] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.439919] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.439919] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.440454] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6017b9-0834-472c-9757-a2261c8e461f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.449097] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab24682-023c-4c61-abaa-307f1e1e46f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.543632] env[63197]: ERROR nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. [ 763.543632] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 763.543632] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.543632] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 763.543632] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.543632] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 763.543632] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.543632] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 763.543632] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.543632] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 763.543632] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.543632] env[63197]: ERROR nova.compute.manager raise self.value [ 763.543632] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.543632] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 763.543632] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.543632] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 763.544240] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.544240] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 763.544240] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. [ 763.544240] env[63197]: ERROR nova.compute.manager [ 763.544240] env[63197]: Traceback (most recent call last): [ 763.544240] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 763.544240] env[63197]: listener.cb(fileno) [ 763.544240] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.544240] env[63197]: result = function(*args, **kwargs) [ 763.544240] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.544240] env[63197]: return func(*args, **kwargs) [ 763.544240] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.544240] env[63197]: raise e [ 763.544240] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.544240] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 763.544240] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.544240] env[63197]: created_port_ids = self._update_ports_for_instance( [ 763.544240] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.544240] env[63197]: with excutils.save_and_reraise_exception(): [ 763.544240] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.544240] env[63197]: self.force_reraise() [ 763.544240] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.544240] env[63197]: raise self.value [ 763.544240] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.544240] env[63197]: updated_port = self._update_port( [ 763.544240] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.544240] env[63197]: _ensure_no_port_binding_failure(port) [ 763.544240] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.544240] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 763.545152] env[63197]: nova.exception.PortBindingFailed: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. [ 763.545152] env[63197]: Removing descriptor: 18 [ 763.545152] env[63197]: ERROR nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Traceback (most recent call last): [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] yield resources [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self.driver.spawn(context, instance, image_meta, [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 763.545152] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] vm_ref = self.build_virtual_machine(instance, [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] for vif in network_info: [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return self._sync_wrapper(fn, *args, **kwargs) [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self.wait() [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self[:] = self._gt.wait() [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return self._exit_event.wait() [ 763.545841] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] result = hub.switch() [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return self.greenlet.switch() [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] result = function(*args, **kwargs) [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return func(*args, **kwargs) [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] raise e [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] nwinfo = self.network_api.allocate_for_instance( [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 763.546287] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] created_port_ids = self._update_ports_for_instance( [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] with excutils.save_and_reraise_exception(): [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self.force_reraise() [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] raise self.value [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] updated_port = self._update_port( [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] _ensure_no_port_binding_failure(port) [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 763.546701] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] raise exception.PortBindingFailed(port_id=port['id']) [ 763.547373] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] nova.exception.PortBindingFailed: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. [ 763.547373] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] [ 763.547373] env[63197]: INFO nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Terminating instance [ 763.547690] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.547848] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.548025] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.630089] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.630321] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "9c1caa2b-c369-425a-8726-cddadf06f338" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.889476] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.901480] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Releasing lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.902110] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 763.903033] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34841904-a84c-4bab-9104-1c64494b00e1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.914153] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ace11c8-fa1f-4e57-b15f-4a6498700b4e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.937945] env[63197]: WARNING nova.virt.vmwareapi.driver [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance e28f14fc-dd09-4baf-9be9-cba280f47f80 could not be found. [ 763.938436] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 763.940784] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82cefe47-7d51-4da0-bfbc-e547aceda75f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.949212] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9346cf-f0a8-411b-9fe7-e2f1e436f3ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.973123] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e28f14fc-dd09-4baf-9be9-cba280f47f80 could not be found. [ 763.973300] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.973521] env[63197]: INFO nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Took 0.07 seconds to destroy the instance on the hypervisor. [ 763.973792] env[63197]: DEBUG oslo.service.loopingcall [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.976265] env[63197]: DEBUG nova.compute.manager [-] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.976375] env[63197]: DEBUG nova.network.neutron [-] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.991389] env[63197]: DEBUG nova.network.neutron [-] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.000018] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.066568] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.177692] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.271789] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6bca38-ffd7-4b4a-a30c-c43cb5b57ef2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.280057] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0867ca43-3a74-4acb-ae9d-b5612789480d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.311773] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34940803-6cff-4423-9c9a-e7654dff2c8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.319511] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f02c5bd-0454-4750-8d6f-6f8d40e324a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.334487] env[63197]: DEBUG nova.compute.provider_tree [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.494373] env[63197]: DEBUG nova.network.neutron [-] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.503990] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Releasing lock "refresh_cache-715c53a9-224e-4457-9df9-a2ada9fcae93" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.503990] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 764.504179] env[63197]: DEBUG nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.504179] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.529068] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.679922] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.680279] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.680471] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.680753] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5190bc44-dfe1-4d3c-aadb-85707a099247 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.689829] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbf8607-65d4-43c9-8b61-c7e267cc61f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.710690] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c3137a4-4e51-4bf9-b7d3-779de61b69b9 could not be found. [ 764.710894] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.711107] env[63197]: INFO nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 764.711350] env[63197]: DEBUG oslo.service.loopingcall [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.711562] env[63197]: DEBUG nova.compute.manager [-] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.711657] env[63197]: DEBUG nova.network.neutron [-] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.726104] env[63197]: DEBUG nova.network.neutron [-] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.837845] env[63197]: DEBUG nova.scheduler.client.report [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.997605] env[63197]: INFO nova.compute.manager [-] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Took 1.02 seconds to deallocate network for instance. [ 765.031275] env[63197]: DEBUG nova.network.neutron [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.075113] env[63197]: DEBUG nova.compute.manager [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Received event network-changed-e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 765.075305] env[63197]: DEBUG nova.compute.manager [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Refreshing instance network info cache due to event network-changed-e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 765.075526] env[63197]: DEBUG oslo_concurrency.lockutils [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] Acquiring lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.075667] env[63197]: DEBUG oslo_concurrency.lockutils [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] Acquired lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.075824] env[63197]: DEBUG nova.network.neutron [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Refreshing network info cache for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 765.230030] env[63197]: DEBUG nova.network.neutron [-] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.343192] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.343733] env[63197]: ERROR nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Traceback (most recent call last): [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self.driver.spawn(context, instance, image_meta, [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self._vmops.spawn(context, instance, image_meta, injected_files, [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] vm_ref = self.build_virtual_machine(instance, [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] vif_infos = vmwarevif.get_vif_info(self._session, [ 765.343733] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] for vif in network_info: [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return self._sync_wrapper(fn, *args, **kwargs) [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self.wait() [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self[:] = self._gt.wait() [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return self._exit_event.wait() [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] result = hub.switch() [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 765.344114] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return self.greenlet.switch() [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] result = function(*args, **kwargs) [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] return func(*args, **kwargs) [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] raise e [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] nwinfo = self.network_api.allocate_for_instance( [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] created_port_ids = self._update_ports_for_instance( [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] with excutils.save_and_reraise_exception(): [ 765.344462] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] self.force_reraise() [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] raise self.value [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] updated_port = self._update_port( [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] _ensure_no_port_binding_failure(port) [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] raise exception.PortBindingFailed(port_id=port['id']) [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] nova.exception.PortBindingFailed: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. [ 765.345112] env[63197]: ERROR nova.compute.manager [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] [ 765.345419] env[63197]: DEBUG nova.compute.utils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 765.345809] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.592s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.348941] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Build of instance 86a728c9-27d3-4c98-a4d6-6b0736f24648 was re-scheduled: Binding failed for port 5fb83a4b-dfb7-444d-9a9b-a3c24cbe5524, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 765.348941] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 765.349123] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquiring lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.349756] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Acquired lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.349756] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.533666] env[63197]: INFO nova.compute.manager [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] [instance: 715c53a9-224e-4457-9df9-a2ada9fcae93] Took 1.03 seconds to deallocate network for instance. [ 765.558929] env[63197]: INFO nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Took 0.56 seconds to detach 1 volumes for instance. [ 765.559312] env[63197]: DEBUG nova.compute.claims [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 765.559672] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.593284] env[63197]: DEBUG nova.network.neutron [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.662349] env[63197]: DEBUG nova.network.neutron [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.731532] env[63197]: INFO nova.compute.manager [-] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Took 1.02 seconds to deallocate network for instance. [ 765.734279] env[63197]: DEBUG nova.compute.claims [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 765.734279] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.870486] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.950416] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.165384] env[63197]: DEBUG oslo_concurrency.lockutils [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] Releasing lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.165649] env[63197]: DEBUG nova.compute.manager [req-61f70a09-db7a-4a20-be44-e2f730e38369 req-ba8d3185-9e12-434f-a5db-c3a5395b450a service nova] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Received event network-vif-deleted-e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.167149] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8313ef-219b-4c0e-b3de-23a693d5432c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.174542] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddf316a-38b1-4900-ad76-fb498ea94376 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.202908] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb9ee4c-266f-4d6b-9d75-e78b79de45e0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.209979] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f090dc52-ded4-484c-be36-41fd94068cf7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.223760] env[63197]: DEBUG nova.compute.provider_tree [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.454402] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Releasing lock "refresh_cache-86a728c9-27d3-4c98-a4d6-6b0736f24648" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.454670] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 766.454853] env[63197]: DEBUG nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 766.456973] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.469478] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.563417] env[63197]: INFO nova.scheduler.client.report [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Deleted allocations for instance 715c53a9-224e-4457-9df9-a2ada9fcae93 [ 766.727209] env[63197]: DEBUG nova.scheduler.client.report [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.975695] env[63197]: DEBUG nova.network.neutron [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.073569] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d5d7145c-fc06-4761-b050-9b79ba3c358c tempest-ImagesNegativeTestJSON-1765356741 tempest-ImagesNegativeTestJSON-1765356741-project-member] Lock "715c53a9-224e-4457-9df9-a2ada9fcae93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.084s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.232106] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.886s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.232755] env[63197]: ERROR nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Traceback (most recent call last): [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self.driver.spawn(context, instance, image_meta, [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] vm_ref = self.build_virtual_machine(instance, [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] vif_infos = vmwarevif.get_vif_info(self._session, [ 767.232755] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] for vif in network_info: [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return self._sync_wrapper(fn, *args, **kwargs) [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self.wait() [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self[:] = self._gt.wait() [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return self._exit_event.wait() [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] result = hub.switch() [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 767.233177] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return self.greenlet.switch() [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] result = function(*args, **kwargs) [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] return func(*args, **kwargs) [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] raise e [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] nwinfo = self.network_api.allocate_for_instance( [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] created_port_ids = self._update_ports_for_instance( [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] with excutils.save_and_reraise_exception(): [ 767.233604] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] self.force_reraise() [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] raise self.value [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] updated_port = self._update_port( [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] _ensure_no_port_binding_failure(port) [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] raise exception.PortBindingFailed(port_id=port['id']) [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] nova.exception.PortBindingFailed: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. [ 767.234013] env[63197]: ERROR nova.compute.manager [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] [ 767.234377] env[63197]: DEBUG nova.compute.utils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 767.234784] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.817s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.236399] env[63197]: INFO nova.compute.claims [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.239055] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Build of instance 218ad539-e2d0-4d4b-9857-ec9d2af0be03 was re-scheduled: Binding failed for port 8506b4cc-4a70-4cb7-b1dc-d10ee8586fe9, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 767.239450] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 767.239674] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Acquiring lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.239818] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Acquired lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.239972] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.478677] env[63197]: INFO nova.compute.manager [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] [instance: 86a728c9-27d3-4c98-a4d6-6b0736f24648] Took 1.02 seconds to deallocate network for instance. [ 767.575651] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 767.762616] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.859711] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.107707] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.362297] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Releasing lock "refresh_cache-218ad539-e2d0-4d4b-9857-ec9d2af0be03" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.362579] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 768.362713] env[63197]: DEBUG nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.362879] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.379338] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.519208] env[63197]: INFO nova.scheduler.client.report [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Deleted allocations for instance 86a728c9-27d3-4c98-a4d6-6b0736f24648 [ 768.605907] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9364655c-6729-4518-a28b-9cdc3e22c904 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.618951] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99167c1-e6c0-43b3-976c-0954ba8bf250 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.647874] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c77431-4761-4f4e-87c8-baeab267e216 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.655115] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf41d664-8e89-4f10-bdd1-1d8f3c30d695 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.668630] env[63197]: DEBUG nova.compute.provider_tree [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.884219] env[63197]: DEBUG nova.network.neutron [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.025180] env[63197]: DEBUG oslo_concurrency.lockutils [None req-277d2a90-7333-4cdb-9347-f460e5dc078e tempest-ServerRescueNegativeTestJSON-2025070609 tempest-ServerRescueNegativeTestJSON-2025070609-project-member] Lock "86a728c9-27d3-4c98-a4d6-6b0736f24648" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.979s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.172670] env[63197]: DEBUG nova.scheduler.client.report [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.386706] env[63197]: INFO nova.compute.manager [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] [instance: 218ad539-e2d0-4d4b-9857-ec9d2af0be03] Took 1.02 seconds to deallocate network for instance. [ 769.532485] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.691228] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.691228] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 769.698630] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.490s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.060208] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.204992] env[63197]: DEBUG nova.compute.utils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.209284] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 770.209391] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 770.261612] env[63197]: DEBUG nova.policy [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccf7a4049c9d4341a59b7a050a1d5f09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d78a0bec6e64f7dacabbf83f0b1b13f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 770.425630] env[63197]: INFO nova.scheduler.client.report [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Deleted allocations for instance 218ad539-e2d0-4d4b-9857-ec9d2af0be03 [ 770.587173] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d8eaf76-73f4-46cc-8dfd-fd05e51109b2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.592777] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca01e65-7adb-43f6-a4c6-4991c47cbc06 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.629446] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443f60a3-d98a-4b90-9035-2f314299117b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.638343] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d85940-3495-430e-a3bc-4717f391a564 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.656123] env[63197]: DEBUG nova.compute.provider_tree [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.710962] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 770.766352] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Successfully created port: e0752ff9-b6da-4e0a-9644-5d4455e065ed {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.937025] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d344317f-e274-492a-8b46-0d2f2b76b419 tempest-ServerRescueTestJSONUnderV235-1345172076 tempest-ServerRescueTestJSONUnderV235-1345172076-project-member] Lock "218ad539-e2d0-4d4b-9857-ec9d2af0be03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.415s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.160080] env[63197]: DEBUG nova.scheduler.client.report [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.441789] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.665825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.666483] env[63197]: ERROR nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Traceback (most recent call last): [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self.driver.spawn(context, instance, image_meta, [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] vm_ref = self.build_virtual_machine(instance, [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] vif_infos = vmwarevif.get_vif_info(self._session, [ 771.666483] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] for vif in network_info: [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] return self._sync_wrapper(fn, *args, **kwargs) [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self.wait() [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self[:] = self._gt.wait() [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] return self._exit_event.wait() [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] current.throw(*self._exc) [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 771.666845] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] result = function(*args, **kwargs) [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] return func(*args, **kwargs) [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] raise e [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] nwinfo = self.network_api.allocate_for_instance( [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] created_port_ids = self._update_ports_for_instance( [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] with excutils.save_and_reraise_exception(): [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] self.force_reraise() [ 771.667225] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] raise self.value [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] updated_port = self._update_port( [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] _ensure_no_port_binding_failure(port) [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] raise exception.PortBindingFailed(port_id=port['id']) [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] nova.exception.PortBindingFailed: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. [ 771.667578] env[63197]: ERROR nova.compute.manager [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] [ 771.667578] env[63197]: DEBUG nova.compute.utils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 771.670049] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.826s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.670471] env[63197]: INFO nova.compute.claims [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.673282] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Build of instance d9f1dcc1-77fd-4c8c-b930-eb421691dc94 was re-scheduled: Binding failed for port 884ecdf1-7984-42fc-a266-49330148a117, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 771.673790] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 771.674053] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Acquiring lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.674234] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Acquired lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.674425] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.730380] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 771.764682] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.765169] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.765485] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.765796] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.766101] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.766637] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.766998] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.767313] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.767635] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.768075] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.768558] env[63197]: DEBUG nova.virt.hardware [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.770934] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4e95b5-eb23-4aaa-a63d-c2cb2c36d834 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.781742] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43c9024-5668-4b44-b178-8851cddcc0d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.969186] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.127640] env[63197]: DEBUG nova.compute.manager [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Received event network-changed-e0752ff9-b6da-4e0a-9644-5d4455e065ed {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.128014] env[63197]: DEBUG nova.compute.manager [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Refreshing instance network info cache due to event network-changed-e0752ff9-b6da-4e0a-9644-5d4455e065ed. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.128684] env[63197]: DEBUG oslo_concurrency.lockutils [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] Acquiring lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.128986] env[63197]: DEBUG oslo_concurrency.lockutils [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] Acquired lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.129335] env[63197]: DEBUG nova.network.neutron [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Refreshing network info cache for port e0752ff9-b6da-4e0a-9644-5d4455e065ed {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.200785] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.367541] env[63197]: ERROR nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. [ 772.367541] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 772.367541] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.367541] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 772.367541] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.367541] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 772.367541] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.367541] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 772.367541] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.367541] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 772.367541] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.367541] env[63197]: ERROR nova.compute.manager raise self.value [ 772.367541] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.367541] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 772.367541] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.367541] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 772.368099] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.368099] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 772.368099] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. [ 772.368099] env[63197]: ERROR nova.compute.manager [ 772.368099] env[63197]: Traceback (most recent call last): [ 772.368099] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 772.368099] env[63197]: listener.cb(fileno) [ 772.368099] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.368099] env[63197]: result = function(*args, **kwargs) [ 772.368099] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.368099] env[63197]: return func(*args, **kwargs) [ 772.368099] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.368099] env[63197]: raise e [ 772.368099] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.368099] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 772.368099] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.368099] env[63197]: created_port_ids = self._update_ports_for_instance( [ 772.368099] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.368099] env[63197]: with excutils.save_and_reraise_exception(): [ 772.368099] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.368099] env[63197]: self.force_reraise() [ 772.368099] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.368099] env[63197]: raise self.value [ 772.368099] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.368099] env[63197]: updated_port = self._update_port( [ 772.368099] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.368099] env[63197]: _ensure_no_port_binding_failure(port) [ 772.368099] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.368099] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 772.368947] env[63197]: nova.exception.PortBindingFailed: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. [ 772.368947] env[63197]: Removing descriptor: 18 [ 772.368947] env[63197]: ERROR nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Traceback (most recent call last): [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] yield resources [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self.driver.spawn(context, instance, image_meta, [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.368947] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] vm_ref = self.build_virtual_machine(instance, [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] for vif in network_info: [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return self._sync_wrapper(fn, *args, **kwargs) [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self.wait() [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self[:] = self._gt.wait() [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return self._exit_event.wait() [ 772.369356] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] result = hub.switch() [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return self.greenlet.switch() [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] result = function(*args, **kwargs) [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return func(*args, **kwargs) [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] raise e [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] nwinfo = self.network_api.allocate_for_instance( [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.369782] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] created_port_ids = self._update_ports_for_instance( [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] with excutils.save_and_reraise_exception(): [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self.force_reraise() [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] raise self.value [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] updated_port = self._update_port( [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] _ensure_no_port_binding_failure(port) [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.370418] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] raise exception.PortBindingFailed(port_id=port['id']) [ 772.370804] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] nova.exception.PortBindingFailed: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. [ 772.370804] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] [ 772.370804] env[63197]: INFO nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Terminating instance [ 772.371248] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.506199] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.658344] env[63197]: DEBUG nova.network.neutron [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.790529] env[63197]: DEBUG nova.network.neutron [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.009128] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Releasing lock "refresh_cache-d9f1dcc1-77fd-4c8c-b930-eb421691dc94" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.009371] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 773.009579] env[63197]: DEBUG nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.009770] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 773.034719] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.113516] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576d504f-df6a-4cc2-8b59-5a8d14966920 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.122874] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0fec4a-7b22-4242-8151-0793d2d29ce2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.159399] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40cc14f-ffda-4ff2-ad7c-8e2a86df9bc6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.168633] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec9e256-b98e-4b5f-a64e-d41fc7cccc01 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.183476] env[63197]: DEBUG nova.compute.provider_tree [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.299705] env[63197]: DEBUG oslo_concurrency.lockutils [req-05a4ee1e-9620-478d-860f-cf30460b15dd req-4f825656-ef10-45b7-a915-4e3cea1ca73c service nova] Releasing lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.299705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.299705] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.539484] env[63197]: DEBUG nova.network.neutron [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.689491] env[63197]: DEBUG nova.scheduler.client.report [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.824895] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.899272] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.042710] env[63197]: INFO nova.compute.manager [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] [instance: d9f1dcc1-77fd-4c8c-b930-eb421691dc94] Took 1.03 seconds to deallocate network for instance. [ 774.160721] env[63197]: DEBUG nova.compute.manager [req-30ae3460-9c7b-4800-acb7-a017f2ec1de1 req-b45c7330-8243-496f-931e-f9ee6f8f7115 service nova] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Received event network-vif-deleted-e0752ff9-b6da-4e0a-9644-5d4455e065ed {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.194295] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.194798] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.197649] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.171s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.388799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "72f9b0c4-69b9-49f2-8665-ff77151883af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.389050] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.403167] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.403167] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.403167] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.403167] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36fa85fb-85ee-463c-af86-24643b66aab6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.412402] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2e98da-95b7-4820-8680-0b52903d6c48 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.433156] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15120438-f2a8-4295-a70f-6ec8afdd9030 could not be found. [ 774.434147] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.434147] env[63197]: INFO nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Took 0.03 seconds to destroy the instance on the hypervisor. [ 774.434147] env[63197]: DEBUG oslo.service.loopingcall [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.434147] env[63197]: DEBUG nova.compute.manager [-] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.434147] env[63197]: DEBUG nova.network.neutron [-] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.447026] env[63197]: DEBUG nova.network.neutron [-] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.701886] env[63197]: DEBUG nova.compute.utils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.706775] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.707076] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 774.765405] env[63197]: DEBUG nova.policy [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6380f852eb614c4fb8a82872d6bf97c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '104c325de57847fd9d55ccdb10b8af97', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.949416] env[63197]: DEBUG nova.network.neutron [-] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.067695] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db709ca6-8ed9-4a11-8411-e9e29e685abe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.075681] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43920361-17b4-4095-af61-fac655286133 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.079915] env[63197]: INFO nova.scheduler.client.report [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Deleted allocations for instance d9f1dcc1-77fd-4c8c-b930-eb421691dc94 [ 775.117891] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Successfully created port: 336f3ea4-949f-4784-9335-47dbd7681814 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 775.120730] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2144ac43-4a53-49d4-bfb8-fd8940be869f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.128784] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ad3b54-87ae-4aaf-8891-c2c7353b2bed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.143692] env[63197]: DEBUG nova.compute.provider_tree [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.207729] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.452131] env[63197]: INFO nova.compute.manager [-] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Took 1.02 seconds to deallocate network for instance. [ 775.454532] env[63197]: DEBUG nova.compute.claims [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 775.454713] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.588605] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3b63a48-af11-4aaf-95e5-ad7ebe2ed93d tempest-ImagesOneServerNegativeTestJSON-467385745 tempest-ImagesOneServerNegativeTestJSON-467385745-project-member] Lock "d9f1dcc1-77fd-4c8c-b930-eb421691dc94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.054s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.648073] env[63197]: DEBUG nova.scheduler.client.report [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.881446] env[63197]: DEBUG nova.compute.manager [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Received event network-changed-336f3ea4-949f-4784-9335-47dbd7681814 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.881851] env[63197]: DEBUG nova.compute.manager [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Refreshing instance network info cache due to event network-changed-336f3ea4-949f-4784-9335-47dbd7681814. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 775.882182] env[63197]: DEBUG oslo_concurrency.lockutils [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] Acquiring lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.882366] env[63197]: DEBUG oslo_concurrency.lockutils [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] Acquired lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.882705] env[63197]: DEBUG nova.network.neutron [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Refreshing network info cache for port 336f3ea4-949f-4784-9335-47dbd7681814 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.096128] env[63197]: ERROR nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. [ 776.096128] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 776.096128] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.096128] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 776.096128] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.096128] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 776.096128] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.096128] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 776.096128] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.096128] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 776.096128] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.096128] env[63197]: ERROR nova.compute.manager raise self.value [ 776.096128] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.096128] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 776.096128] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.096128] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 776.096613] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.096613] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 776.096613] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. [ 776.096613] env[63197]: ERROR nova.compute.manager [ 776.096613] env[63197]: Traceback (most recent call last): [ 776.096613] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 776.096613] env[63197]: listener.cb(fileno) [ 776.096613] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.096613] env[63197]: result = function(*args, **kwargs) [ 776.096613] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.096613] env[63197]: return func(*args, **kwargs) [ 776.096613] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.096613] env[63197]: raise e [ 776.096613] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.096613] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 776.096613] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.096613] env[63197]: created_port_ids = self._update_ports_for_instance( [ 776.096613] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.096613] env[63197]: with excutils.save_and_reraise_exception(): [ 776.096613] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.096613] env[63197]: self.force_reraise() [ 776.096613] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.096613] env[63197]: raise self.value [ 776.096613] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.096613] env[63197]: updated_port = self._update_port( [ 776.096613] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.096613] env[63197]: _ensure_no_port_binding_failure(port) [ 776.096613] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.096613] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 776.097573] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. [ 776.097573] env[63197]: Removing descriptor: 19 [ 776.097573] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 776.152952] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.153610] env[63197]: ERROR nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Traceback (most recent call last): [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self.driver.spawn(context, instance, image_meta, [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] vm_ref = self.build_virtual_machine(instance, [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.153610] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] for vif in network_info: [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return self._sync_wrapper(fn, *args, **kwargs) [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self.wait() [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self[:] = self._gt.wait() [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return self._exit_event.wait() [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] result = hub.switch() [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 776.154025] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return self.greenlet.switch() [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] result = function(*args, **kwargs) [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] return func(*args, **kwargs) [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] raise e [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] nwinfo = self.network_api.allocate_for_instance( [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] created_port_ids = self._update_ports_for_instance( [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] with excutils.save_and_reraise_exception(): [ 776.154394] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] self.force_reraise() [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] raise self.value [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] updated_port = self._update_port( [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] _ensure_no_port_binding_failure(port) [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] raise exception.PortBindingFailed(port_id=port['id']) [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] nova.exception.PortBindingFailed: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. [ 776.154748] env[63197]: ERROR nova.compute.manager [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] [ 776.155064] env[63197]: DEBUG nova.compute.utils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.155414] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.850s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.157202] env[63197]: INFO nova.compute.claims [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.160383] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Build of instance a1ba5772-9e73-4a51-ae62-d9c17a64ff0d was re-scheduled: Binding failed for port 5afd3f1b-0841-41b6-b92e-10497d48f1e6, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 776.160878] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 776.161233] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Acquiring lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.161445] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Acquired lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.161658] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.218665] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.244539] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.244797] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.245836] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.245836] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.245836] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.245836] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.245836] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.246055] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.246055] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.246132] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.246266] env[63197]: DEBUG nova.virt.hardware [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.247128] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c675d952-def8-4615-89db-90724b5677ca {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.257352] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b496fde-cf85-4716-9db6-b6c564b8d9f8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.268483] env[63197]: ERROR nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Traceback (most recent call last): [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] yield resources [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self.driver.spawn(context, instance, image_meta, [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self._vmops.spawn(context, instance, image_meta, injected_files, [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] vm_ref = self.build_virtual_machine(instance, [ 776.268483] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] vif_infos = vmwarevif.get_vif_info(self._session, [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] for vif in network_info: [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] return self._sync_wrapper(fn, *args, **kwargs) [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self.wait() [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self[:] = self._gt.wait() [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] return self._exit_event.wait() [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 776.268924] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] current.throw(*self._exc) [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] result = function(*args, **kwargs) [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] return func(*args, **kwargs) [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] raise e [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] nwinfo = self.network_api.allocate_for_instance( [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] created_port_ids = self._update_ports_for_instance( [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] with excutils.save_and_reraise_exception(): [ 776.269360] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self.force_reraise() [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] raise self.value [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] updated_port = self._update_port( [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] _ensure_no_port_binding_failure(port) [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] raise exception.PortBindingFailed(port_id=port['id']) [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] nova.exception.PortBindingFailed: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. [ 776.269722] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] [ 776.269722] env[63197]: INFO nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Terminating instance [ 776.270674] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.402555] env[63197]: DEBUG nova.network.neutron [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.501231] env[63197]: DEBUG nova.network.neutron [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.620610] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.684043] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.791786] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.007690] env[63197]: DEBUG oslo_concurrency.lockutils [req-96cccc02-a9a4-4174-add5-c81d9c1ad10c req-a6b77ffa-b763-4b9e-8fcf-56851de6ca4e service nova] Releasing lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.008349] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquired lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.008349] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.292873] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Releasing lock "refresh_cache-a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.293109] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 777.293321] env[63197]: DEBUG nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.293486] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 777.311132] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.488795] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995426bc-f9e7-423b-a903-d2031bb62b99 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.496397] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65aa6a0-a66d-4c83-af2d-c18fe77fc62b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.529601] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f394fda5-14a3-46d3-a900-bc1b89af9e65 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.540024] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0a098b-a472-4f11-8d95-085f122c946d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.550806] env[63197]: DEBUG nova.compute.provider_tree [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.552791] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.643108] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.813728] env[63197]: DEBUG nova.network.neutron [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.904297] env[63197]: DEBUG nova.compute.manager [req-fc4674da-4cd9-4829-b659-cf68e98330cb req-67ddf222-4c77-4d13-a760-3267ee2c567a service nova] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Received event network-vif-deleted-336f3ea4-949f-4784-9335-47dbd7681814 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.055985] env[63197]: DEBUG nova.scheduler.client.report [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 778.145997] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Releasing lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.146155] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 778.146353] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 778.146635] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-690aa063-3b26-4bb8-9ad1-9d145f6dbf94 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.155646] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e7ee4f-1d86-4f2b-8525-759b26abc959 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.176485] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b5935562-9a77-4b92-ac73-a7af496ccced could not be found. [ 778.176682] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 778.176856] env[63197]: INFO nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Took 0.03 seconds to destroy the instance on the hypervisor. [ 778.177098] env[63197]: DEBUG oslo.service.loopingcall [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.177458] env[63197]: DEBUG nova.compute.manager [-] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.177555] env[63197]: DEBUG nova.network.neutron [-] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.191965] env[63197]: DEBUG nova.network.neutron [-] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.316294] env[63197]: INFO nova.compute.manager [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] [instance: a1ba5772-9e73-4a51-ae62-d9c17a64ff0d] Took 1.02 seconds to deallocate network for instance. [ 778.562080] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.562080] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 778.564554] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.005s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.694511] env[63197]: DEBUG nova.network.neutron [-] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.070113] env[63197]: DEBUG nova.compute.utils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.075295] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.075463] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.121592] env[63197]: DEBUG nova.policy [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df6c10bca4144b2db6a0ee47539966d5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60d59853e04d462286952d221b398c66', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.197191] env[63197]: INFO nova.compute.manager [-] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Took 1.02 seconds to deallocate network for instance. [ 779.199539] env[63197]: DEBUG nova.compute.claims [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 779.199717] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.351425] env[63197]: INFO nova.scheduler.client.report [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Deleted allocations for instance a1ba5772-9e73-4a51-ae62-d9c17a64ff0d [ 779.390407] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9d71f4-61e3-439f-93e4-5a20d91d650a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.398873] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cecb0912-51f1-475c-8644-a688f990ff68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.430115] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Successfully created port: 393667ba-b110-418e-bee0-c07aed3c8a5b {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 779.432448] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0e1acf-4498-4287-840e-f46339afa6ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.439757] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a56fa4-9926-4c8c-b351-e53d53132fbc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.452651] env[63197]: DEBUG nova.compute.provider_tree [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.576603] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.861747] env[63197]: DEBUG oslo_concurrency.lockutils [None req-241f789d-9f37-4994-b98e-3ee1df7cd630 tempest-InstanceActionsV221TestJSON-1657060385 tempest-InstanceActionsV221TestJSON-1657060385-project-member] Lock "a1ba5772-9e73-4a51-ae62-d9c17a64ff0d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.322s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.956247] env[63197]: DEBUG nova.scheduler.client.report [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 780.214067] env[63197]: DEBUG nova.compute.manager [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Received event network-changed-393667ba-b110-418e-bee0-c07aed3c8a5b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.214505] env[63197]: DEBUG nova.compute.manager [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Refreshing instance network info cache due to event network-changed-393667ba-b110-418e-bee0-c07aed3c8a5b. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 780.214505] env[63197]: DEBUG oslo_concurrency.lockutils [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] Acquiring lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.214651] env[63197]: DEBUG oslo_concurrency.lockutils [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] Acquired lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.214764] env[63197]: DEBUG nova.network.neutron [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Refreshing network info cache for port 393667ba-b110-418e-bee0-c07aed3c8a5b {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 780.365426] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.396580] env[63197]: ERROR nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. [ 780.396580] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 780.396580] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.396580] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 780.396580] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.396580] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 780.396580] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.396580] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 780.396580] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.396580] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 780.396580] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.396580] env[63197]: ERROR nova.compute.manager raise self.value [ 780.396580] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.396580] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 780.396580] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.396580] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 780.397400] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.397400] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 780.397400] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. [ 780.397400] env[63197]: ERROR nova.compute.manager [ 780.397400] env[63197]: Traceback (most recent call last): [ 780.397400] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 780.397400] env[63197]: listener.cb(fileno) [ 780.397400] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.397400] env[63197]: result = function(*args, **kwargs) [ 780.397400] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 780.397400] env[63197]: return func(*args, **kwargs) [ 780.397400] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.397400] env[63197]: raise e [ 780.397400] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.397400] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 780.397400] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.397400] env[63197]: created_port_ids = self._update_ports_for_instance( [ 780.397400] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.397400] env[63197]: with excutils.save_and_reraise_exception(): [ 780.397400] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.397400] env[63197]: self.force_reraise() [ 780.397400] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.397400] env[63197]: raise self.value [ 780.397400] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.397400] env[63197]: updated_port = self._update_port( [ 780.397400] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.397400] env[63197]: _ensure_no_port_binding_failure(port) [ 780.397400] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.397400] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 780.398800] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. [ 780.398800] env[63197]: Removing descriptor: 19 [ 780.461950] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.897s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.462626] env[63197]: ERROR nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Traceback (most recent call last): [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self.driver.spawn(context, instance, image_meta, [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] vm_ref = self.build_virtual_machine(instance, [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.462626] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] for vif in network_info: [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] return self._sync_wrapper(fn, *args, **kwargs) [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self.wait() [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self[:] = self._gt.wait() [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] return self._exit_event.wait() [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] current.throw(*self._exc) [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.463041] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] result = function(*args, **kwargs) [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] return func(*args, **kwargs) [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] raise e [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] nwinfo = self.network_api.allocate_for_instance( [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] created_port_ids = self._update_ports_for_instance( [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] with excutils.save_and_reraise_exception(): [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] self.force_reraise() [ 780.463465] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] raise self.value [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] updated_port = self._update_port( [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] _ensure_no_port_binding_failure(port) [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] raise exception.PortBindingFailed(port_id=port['id']) [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] nova.exception.PortBindingFailed: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. [ 780.463871] env[63197]: ERROR nova.compute.manager [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] [ 780.463871] env[63197]: DEBUG nova.compute.utils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.464472] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.730s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.467490] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Build of instance e28f14fc-dd09-4baf-9be9-cba280f47f80 was re-scheduled: Binding failed for port b936185c-ff52-4611-8304-2d440945a78f, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 780.467903] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 780.468199] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Acquiring lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.468294] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Acquired lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.469549] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.589884] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 780.618135] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.618467] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.618657] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.618933] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.619102] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.619311] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.619558] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.619748] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.619943] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.620169] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.620436] env[63197]: DEBUG nova.virt.hardware [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.621415] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea969f5e-723b-4ba1-b7fd-5873c05b884f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.630208] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044ff859-f4d3-4321-b5e4-3b9947acdc14 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.645153] env[63197]: ERROR nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Traceback (most recent call last): [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] yield resources [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self.driver.spawn(context, instance, image_meta, [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] vm_ref = self.build_virtual_machine(instance, [ 780.645153] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] for vif in network_info: [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] return self._sync_wrapper(fn, *args, **kwargs) [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self.wait() [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self[:] = self._gt.wait() [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] return self._exit_event.wait() [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 780.645598] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] current.throw(*self._exc) [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] result = function(*args, **kwargs) [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] return func(*args, **kwargs) [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] raise e [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] nwinfo = self.network_api.allocate_for_instance( [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] created_port_ids = self._update_ports_for_instance( [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] with excutils.save_and_reraise_exception(): [ 780.646011] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self.force_reraise() [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] raise self.value [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] updated_port = self._update_port( [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] _ensure_no_port_binding_failure(port) [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] raise exception.PortBindingFailed(port_id=port['id']) [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] nova.exception.PortBindingFailed: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. [ 780.646433] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] [ 780.646433] env[63197]: INFO nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Terminating instance [ 780.647719] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Acquiring lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.736151] env[63197]: DEBUG nova.network.neutron [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.891521] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.001862] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.017300] env[63197]: DEBUG nova.network.neutron [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.094189] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.317035] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2e2cd5-233a-4c9b-83e6-d675b12d2de4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.325892] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7337840-c7f7-4cb0-9d5a-fc781fee6cdf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.359948] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e80d7d-6672-4280-85ee-60bf56cfff50 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.367236] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0bf443c-7b67-4d90-887c-17c8deb5d0d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.381327] env[63197]: DEBUG nova.compute.provider_tree [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.520876] env[63197]: DEBUG oslo_concurrency.lockutils [req-27b662ad-0580-4e2c-b93f-c583dddab9e3 req-8353f886-9106-487c-9aeb-058550461d76 service nova] Releasing lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.521686] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Acquired lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.521871] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.596775] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Releasing lock "refresh_cache-e28f14fc-dd09-4baf-9be9-cba280f47f80" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.597029] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 781.597211] env[63197]: DEBUG nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 781.597373] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 781.613141] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.888066] env[63197]: DEBUG nova.scheduler.client.report [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 782.039231] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.116226] env[63197]: DEBUG nova.network.neutron [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.127298] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.242897] env[63197]: DEBUG nova.compute.manager [req-05958b95-461c-412d-839a-3c02c258e999 req-21a59be0-189e-4d93-8c06-5beb32164478 service nova] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Received event network-vif-deleted-393667ba-b110-418e-bee0-c07aed3c8a5b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.393648] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.394330] env[63197]: ERROR nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Traceback (most recent call last): [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self.driver.spawn(context, instance, image_meta, [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] vm_ref = self.build_virtual_machine(instance, [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.394330] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] for vif in network_info: [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return self._sync_wrapper(fn, *args, **kwargs) [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self.wait() [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self[:] = self._gt.wait() [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return self._exit_event.wait() [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] result = hub.switch() [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 782.394768] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return self.greenlet.switch() [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] result = function(*args, **kwargs) [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] return func(*args, **kwargs) [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] raise e [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] nwinfo = self.network_api.allocate_for_instance( [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] created_port_ids = self._update_ports_for_instance( [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] with excutils.save_and_reraise_exception(): [ 782.395248] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] self.force_reraise() [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] raise self.value [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] updated_port = self._update_port( [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] _ensure_no_port_binding_failure(port) [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] raise exception.PortBindingFailed(port_id=port['id']) [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] nova.exception.PortBindingFailed: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. [ 782.395659] env[63197]: ERROR nova.compute.manager [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] [ 782.396028] env[63197]: DEBUG nova.compute.utils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.396636] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.289s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.398391] env[63197]: INFO nova.compute.claims [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.401098] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Build of instance 6c3137a4-4e51-4bf9-b7d3-779de61b69b9 was re-scheduled: Binding failed for port e8bc5b3f-0229-4ffe-a5f3-04234bdaef0d, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 782.401568] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 782.401789] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.401936] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.402116] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.619854] env[63197]: INFO nova.compute.manager [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] [instance: e28f14fc-dd09-4baf-9be9-cba280f47f80] Took 1.02 seconds to deallocate network for instance. [ 782.629470] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Releasing lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.629845] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 782.630045] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 782.630331] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aed13ebc-b89e-42cc-9c28-2c90d4c7882a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.639307] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925d09e1-de4e-45c8-a0cb-8a6d225e3f8e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.661505] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a90e6999-9aab-4b83-98e4-2b71681d1bb5 could not be found. [ 782.661718] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.661895] env[63197]: INFO nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 782.662158] env[63197]: DEBUG oslo.service.loopingcall [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.662375] env[63197]: DEBUG nova.compute.manager [-] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.662468] env[63197]: DEBUG nova.network.neutron [-] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.675246] env[63197]: DEBUG nova.network.neutron [-] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.920790] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.000417] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.177603] env[63197]: DEBUG nova.network.neutron [-] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.503444] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "refresh_cache-6c3137a4-4e51-4bf9-b7d3-779de61b69b9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.503733] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 783.503856] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.504031] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.528205] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.643535] env[63197]: INFO nova.scheduler.client.report [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Deleted allocations for instance e28f14fc-dd09-4baf-9be9-cba280f47f80 [ 783.668923] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3798fbd-960d-44ab-9362-3f4e60cc5e0d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.676981] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d62eb3-0d3d-4794-858a-c38dbdd5f7de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.680546] env[63197]: INFO nova.compute.manager [-] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Took 1.02 seconds to deallocate network for instance. [ 783.682989] env[63197]: DEBUG nova.compute.claims [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 783.683185] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.710212] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31baa002-7bd9-47cf-a921-ad8e1f5aeb94 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.717486] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc23dc5-0e80-4969-a026-4ae6885458d8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.730501] env[63197]: DEBUG nova.compute.provider_tree [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.030842] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.152917] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b1ef7465-71fc-4896-ae84-1d205eefde2a tempest-ServersTestBootFromVolume-1157872990 tempest-ServersTestBootFromVolume-1157872990-project-member] Lock "e28f14fc-dd09-4baf-9be9-cba280f47f80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.609s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.234072] env[63197]: DEBUG nova.scheduler.client.report [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.534322] env[63197]: INFO nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 6c3137a4-4e51-4bf9-b7d3-779de61b69b9] Took 1.03 seconds to deallocate network for instance. [ 784.656100] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.739424] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.739937] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.742653] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.683s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.744352] env[63197]: INFO nova.compute.claims [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.186159] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.249007] env[63197]: DEBUG nova.compute.utils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.253099] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.253272] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 785.313230] env[63197]: DEBUG nova.policy [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd01e2ee5f34248f0890c08c53618db81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf8a2feb7dd94667ad749041bff42fe0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.563212] env[63197]: INFO nova.scheduler.client.report [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleted allocations for instance 6c3137a4-4e51-4bf9-b7d3-779de61b69b9 [ 785.711193] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 785.711193] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 785.756271] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.820178] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Successfully created port: e0ec3f0b-e3c3-457c-b087-79b0ed2f457b {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.071130] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "6c3137a4-4e51-4bf9-b7d3-779de61b69b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.642s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.100093] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bf469e-29c6-44bb-a2de-94671619dd98 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.108425] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c748038-a5a8-49a6-9c73-143167868f45 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.140719] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234d67ad-d799-446b-b3b9-71298de45bf0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.149328] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223d9ac5-4c06-4c6a-b71f-4bfab601dfdc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.162582] env[63197]: DEBUG nova.compute.provider_tree [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.215312] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.215578] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 786.215646] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Rebuilding the list of instances to heal {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 786.574079] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.665781] env[63197]: DEBUG nova.scheduler.client.report [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.679777] env[63197]: DEBUG nova.compute.manager [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Received event network-changed-e0ec3f0b-e3c3-457c-b087-79b0ed2f457b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.679974] env[63197]: DEBUG nova.compute.manager [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Refreshing instance network info cache due to event network-changed-e0ec3f0b-e3c3-457c-b087-79b0ed2f457b. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 786.680208] env[63197]: DEBUG oslo_concurrency.lockutils [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] Acquiring lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.680353] env[63197]: DEBUG oslo_concurrency.lockutils [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] Acquired lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.680781] env[63197]: DEBUG nova.network.neutron [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Refreshing network info cache for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 786.721328] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 786.721620] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 786.721754] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 786.721961] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 786.722183] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 786.722383] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Didn't find any instances for network info cache update. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 786.722611] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.722873] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.723349] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.723349] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.723488] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.723630] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.723763] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 786.723902] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 786.765951] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.800219] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.800538] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.800665] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.800870] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.801072] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.801275] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.801495] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.801691] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.801854] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.802200] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.802406] env[63197]: DEBUG nova.virt.hardware [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.803690] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b41614b-812b-4bd5-a53f-d89388d5a2f3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.811851] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b6f12b-3619-4faf-8c1e-832f560acaa8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.918132] env[63197]: ERROR nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. [ 786.918132] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 786.918132] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.918132] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 786.918132] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.918132] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 786.918132] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.918132] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 786.918132] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.918132] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 786.918132] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.918132] env[63197]: ERROR nova.compute.manager raise self.value [ 786.918132] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.918132] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 786.918132] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.918132] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 786.918842] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.918842] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 786.918842] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. [ 786.918842] env[63197]: ERROR nova.compute.manager [ 786.918842] env[63197]: Traceback (most recent call last): [ 786.918842] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 786.918842] env[63197]: listener.cb(fileno) [ 786.918842] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.918842] env[63197]: result = function(*args, **kwargs) [ 786.918842] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 786.918842] env[63197]: return func(*args, **kwargs) [ 786.918842] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.918842] env[63197]: raise e [ 786.918842] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.918842] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 786.918842] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.918842] env[63197]: created_port_ids = self._update_ports_for_instance( [ 786.918842] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.918842] env[63197]: with excutils.save_and_reraise_exception(): [ 786.918842] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.918842] env[63197]: self.force_reraise() [ 786.918842] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.918842] env[63197]: raise self.value [ 786.918842] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.918842] env[63197]: updated_port = self._update_port( [ 786.918842] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.918842] env[63197]: _ensure_no_port_binding_failure(port) [ 786.918842] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.918842] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 786.920149] env[63197]: nova.exception.PortBindingFailed: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. [ 786.920149] env[63197]: Removing descriptor: 18 [ 786.920149] env[63197]: ERROR nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Traceback (most recent call last): [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] yield resources [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self.driver.spawn(context, instance, image_meta, [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 786.920149] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] vm_ref = self.build_virtual_machine(instance, [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] for vif in network_info: [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return self._sync_wrapper(fn, *args, **kwargs) [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self.wait() [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self[:] = self._gt.wait() [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return self._exit_event.wait() [ 786.920751] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] result = hub.switch() [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return self.greenlet.switch() [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] result = function(*args, **kwargs) [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return func(*args, **kwargs) [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] raise e [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] nwinfo = self.network_api.allocate_for_instance( [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.921363] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] created_port_ids = self._update_ports_for_instance( [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] with excutils.save_and_reraise_exception(): [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self.force_reraise() [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] raise self.value [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] updated_port = self._update_port( [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] _ensure_no_port_binding_failure(port) [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.921929] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] raise exception.PortBindingFailed(port_id=port['id']) [ 786.922537] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] nova.exception.PortBindingFailed: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. [ 786.922537] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] [ 786.922537] env[63197]: INFO nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Terminating instance [ 786.922537] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.100414] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.171675] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.172259] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.174897] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.206s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.176342] env[63197]: INFO nova.compute.claims [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.208149] env[63197]: DEBUG nova.network.neutron [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.227610] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.287796] env[63197]: DEBUG nova.network.neutron [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.680986] env[63197]: DEBUG nova.compute.utils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.686862] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.686862] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 787.772041] env[63197]: DEBUG nova.policy [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd01e2ee5f34248f0890c08c53618db81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf8a2feb7dd94667ad749041bff42fe0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.791029] env[63197]: DEBUG oslo_concurrency.lockutils [req-934651c4-72b1-43e4-a6fb-ae928bba7bfc req-5b763286-62d7-4e87-a8f3-763ffa8e1fc5 service nova] Releasing lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.791029] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquired lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.791299] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.191753] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.273914] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Successfully created port: b7baad22-e600-46ce-8fda-95751fad5c40 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.312956] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.415814] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.491806] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101e1273-7ffc-413d-a201-36a0f5a9e442 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.499878] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ca57a3-ecdb-4b14-a87d-aa89a1adb096 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.532499] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c327608d-36cc-4353-b622-178442301a14 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.539460] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97566b23-13e7-4121-b389-76e8a0e487f9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.551950] env[63197]: DEBUG nova.compute.provider_tree [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.716698] env[63197]: DEBUG nova.compute.manager [req-0ba2f68b-b5c0-4d88-8259-1773d166f343 req-960fdc9e-f615-4eac-8e55-a4ee835ed791 service nova] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Received event network-vif-deleted-e0ec3f0b-e3c3-457c-b087-79b0ed2f457b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.920333] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Releasing lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.920821] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.921093] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.921446] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2059e61d-57a1-4bc2-a960-0f153686ee6d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.930782] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb1e02d-c994-4121-b2c1-057dabbad934 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.953339] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bd771a95-98b1-4fdb-b213-42d693e027f6 could not be found. [ 788.953614] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 788.953818] env[63197]: INFO nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 788.954125] env[63197]: DEBUG oslo.service.loopingcall [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.954376] env[63197]: DEBUG nova.compute.manager [-] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.954479] env[63197]: DEBUG nova.network.neutron [-] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.973601] env[63197]: DEBUG nova.network.neutron [-] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.057809] env[63197]: DEBUG nova.scheduler.client.report [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.200915] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.229389] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.229564] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.229657] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.229770] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.229911] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.230068] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.230274] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.230434] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.230592] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.230748] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.230913] env[63197]: DEBUG nova.virt.hardware [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.231805] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdd97a6-8207-4b66-893a-1aa9d2c439a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.239857] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30c9fe3-7b9c-48ac-b8b7-63d2e38dc79b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.476098] env[63197]: DEBUG nova.network.neutron [-] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.563434] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.563955] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.566491] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.112s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.981906] env[63197]: INFO nova.compute.manager [-] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Took 1.03 seconds to deallocate network for instance. [ 789.984353] env[63197]: DEBUG nova.compute.claims [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 789.984540] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.039091] env[63197]: ERROR nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. [ 790.039091] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 790.039091] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.039091] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 790.039091] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.039091] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 790.039091] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.039091] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 790.039091] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.039091] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 790.039091] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.039091] env[63197]: ERROR nova.compute.manager raise self.value [ 790.039091] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.039091] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 790.039091] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.039091] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 790.039583] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.039583] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 790.039583] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. [ 790.039583] env[63197]: ERROR nova.compute.manager [ 790.039583] env[63197]: Traceback (most recent call last): [ 790.039583] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 790.039583] env[63197]: listener.cb(fileno) [ 790.039583] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.039583] env[63197]: result = function(*args, **kwargs) [ 790.039583] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.039583] env[63197]: return func(*args, **kwargs) [ 790.039583] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.039583] env[63197]: raise e [ 790.039583] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.039583] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 790.039583] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.039583] env[63197]: created_port_ids = self._update_ports_for_instance( [ 790.039583] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.039583] env[63197]: with excutils.save_and_reraise_exception(): [ 790.039583] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.039583] env[63197]: self.force_reraise() [ 790.039583] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.039583] env[63197]: raise self.value [ 790.039583] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.039583] env[63197]: updated_port = self._update_port( [ 790.039583] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.039583] env[63197]: _ensure_no_port_binding_failure(port) [ 790.039583] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.039583] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 790.040385] env[63197]: nova.exception.PortBindingFailed: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. [ 790.040385] env[63197]: Removing descriptor: 18 [ 790.040385] env[63197]: ERROR nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Traceback (most recent call last): [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] yield resources [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self.driver.spawn(context, instance, image_meta, [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.040385] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] vm_ref = self.build_virtual_machine(instance, [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] for vif in network_info: [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return self._sync_wrapper(fn, *args, **kwargs) [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self.wait() [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self[:] = self._gt.wait() [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return self._exit_event.wait() [ 790.040916] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] result = hub.switch() [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return self.greenlet.switch() [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] result = function(*args, **kwargs) [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return func(*args, **kwargs) [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] raise e [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] nwinfo = self.network_api.allocate_for_instance( [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.041372] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] created_port_ids = self._update_ports_for_instance( [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] with excutils.save_and_reraise_exception(): [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self.force_reraise() [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] raise self.value [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] updated_port = self._update_port( [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] _ensure_no_port_binding_failure(port) [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.041745] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] raise exception.PortBindingFailed(port_id=port['id']) [ 790.042746] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] nova.exception.PortBindingFailed: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. [ 790.042746] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] [ 790.042746] env[63197]: INFO nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Terminating instance [ 790.043692] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.044186] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquired lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.044186] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.072069] env[63197]: DEBUG nova.compute.utils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.076335] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.076501] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.171118] env[63197]: DEBUG nova.policy [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd01e2ee5f34248f0890c08c53618db81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf8a2feb7dd94667ad749041bff42fe0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.447321] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3763e805-3727-4e46-b361-0f68c231d13b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.454982] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7043a7ff-7245-46f2-8276-565aef86da94 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.486618] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f443eb-976f-4b2f-b62d-f638f536461b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.494638] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d97104-a924-422a-8da9-f3f9a2f62f78 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.509147] env[63197]: DEBUG nova.compute.provider_tree [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.550088] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Successfully created port: 9a3f1efe-1c8d-4566-b375-1d15b136f104 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.572250] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.579656] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 790.688492] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.744777] env[63197]: DEBUG nova.compute.manager [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Received event network-changed-b7baad22-e600-46ce-8fda-95751fad5c40 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.744969] env[63197]: DEBUG nova.compute.manager [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Refreshing instance network info cache due to event network-changed-b7baad22-e600-46ce-8fda-95751fad5c40. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 790.745567] env[63197]: DEBUG oslo_concurrency.lockutils [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] Acquiring lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.014089] env[63197]: DEBUG nova.scheduler.client.report [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.191510] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Releasing lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.192064] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 791.192305] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.192631] env[63197]: DEBUG oslo_concurrency.lockutils [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] Acquired lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.193268] env[63197]: DEBUG nova.network.neutron [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Refreshing network info cache for port b7baad22-e600-46ce-8fda-95751fad5c40 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.194585] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e42723d2-9ec4-4904-b28b-b26b98048fd7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.204432] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe07ac1-5122-4d2b-8ff5-c8336ce10559 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.227578] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7a2de2b-d14b-437c-93ff-17ee24bd97ee could not be found. [ 791.227787] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.227959] env[63197]: INFO nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 791.228208] env[63197]: DEBUG oslo.service.loopingcall [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.228477] env[63197]: DEBUG nova.compute.manager [-] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.228575] env[63197]: DEBUG nova.network.neutron [-] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.245599] env[63197]: DEBUG nova.network.neutron [-] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.460472] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "a0a40337-d8d7-448b-afff-f6849e9d37a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.460695] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.511027] env[63197]: ERROR nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. [ 791.511027] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 791.511027] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.511027] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 791.511027] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.511027] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 791.511027] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.511027] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 791.511027] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.511027] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 791.511027] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.511027] env[63197]: ERROR nova.compute.manager raise self.value [ 791.511027] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.511027] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 791.511027] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.511027] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 791.511634] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.511634] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 791.511634] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. [ 791.511634] env[63197]: ERROR nova.compute.manager [ 791.511634] env[63197]: Traceback (most recent call last): [ 791.511634] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 791.511634] env[63197]: listener.cb(fileno) [ 791.511634] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.511634] env[63197]: result = function(*args, **kwargs) [ 791.511634] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.511634] env[63197]: return func(*args, **kwargs) [ 791.511634] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.511634] env[63197]: raise e [ 791.511634] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.511634] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 791.511634] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.511634] env[63197]: created_port_ids = self._update_ports_for_instance( [ 791.511634] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.511634] env[63197]: with excutils.save_and_reraise_exception(): [ 791.511634] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.511634] env[63197]: self.force_reraise() [ 791.511634] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.511634] env[63197]: raise self.value [ 791.511634] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.511634] env[63197]: updated_port = self._update_port( [ 791.511634] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.511634] env[63197]: _ensure_no_port_binding_failure(port) [ 791.511634] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.511634] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 791.512593] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. [ 791.512593] env[63197]: Removing descriptor: 18 [ 791.520425] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.520982] env[63197]: ERROR nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Traceback (most recent call last): [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self.driver.spawn(context, instance, image_meta, [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] vm_ref = self.build_virtual_machine(instance, [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.520982] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] for vif in network_info: [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return self._sync_wrapper(fn, *args, **kwargs) [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self.wait() [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self[:] = self._gt.wait() [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return self._exit_event.wait() [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] result = hub.switch() [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 791.521376] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return self.greenlet.switch() [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] result = function(*args, **kwargs) [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] return func(*args, **kwargs) [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] raise e [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] nwinfo = self.network_api.allocate_for_instance( [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] created_port_ids = self._update_ports_for_instance( [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] with excutils.save_and_reraise_exception(): [ 791.521781] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] self.force_reraise() [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] raise self.value [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] updated_port = self._update_port( [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] _ensure_no_port_binding_failure(port) [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] raise exception.PortBindingFailed(port_id=port['id']) [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] nova.exception.PortBindingFailed: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. [ 791.522215] env[63197]: ERROR nova.compute.manager [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] [ 791.522558] env[63197]: DEBUG nova.compute.utils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 791.522840] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.902s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.524563] env[63197]: INFO nova.compute.claims [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.527371] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Build of instance 15120438-f2a8-4295-a70f-6ec8afdd9030 was re-scheduled: Binding failed for port e0752ff9-b6da-4e0a-9644-5d4455e065ed, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 791.527779] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 791.527994] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.528161] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.528318] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.590205] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 791.615926] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.616332] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.616430] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.616742] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.616742] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.616857] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.620086] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.620086] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.620086] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.620086] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.620086] env[63197]: DEBUG nova.virt.hardware [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.620304] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5546483a-df14-42db-8768-74ba9e3b114c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.626709] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8a999f-f08c-4bc5-87b6-97c1c42e44bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.642825] env[63197]: ERROR nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Traceback (most recent call last): [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] yield resources [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self.driver.spawn(context, instance, image_meta, [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] vm_ref = self.build_virtual_machine(instance, [ 791.642825] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] for vif in network_info: [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] return self._sync_wrapper(fn, *args, **kwargs) [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self.wait() [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self[:] = self._gt.wait() [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] return self._exit_event.wait() [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 791.643432] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] current.throw(*self._exc) [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] result = function(*args, **kwargs) [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] return func(*args, **kwargs) [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] raise e [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] nwinfo = self.network_api.allocate_for_instance( [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] created_port_ids = self._update_ports_for_instance( [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] with excutils.save_and_reraise_exception(): [ 791.645564] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self.force_reraise() [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] raise self.value [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] updated_port = self._update_port( [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] _ensure_no_port_binding_failure(port) [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] raise exception.PortBindingFailed(port_id=port['id']) [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] nova.exception.PortBindingFailed: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. [ 791.645981] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] [ 791.645981] env[63197]: INFO nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Terminating instance [ 791.646343] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.646343] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquired lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.646343] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.733330] env[63197]: DEBUG nova.network.neutron [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.747741] env[63197]: DEBUG nova.network.neutron [-] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.826518] env[63197]: DEBUG nova.network.neutron [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.048083] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.118919] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.164915] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.244866] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.250900] env[63197]: INFO nova.compute.manager [-] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Took 1.02 seconds to deallocate network for instance. [ 792.252783] env[63197]: DEBUG nova.compute.claims [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 792.252958] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.329814] env[63197]: DEBUG oslo_concurrency.lockutils [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] Releasing lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.329814] env[63197]: DEBUG nova.compute.manager [req-4c4a2b92-286b-4faf-8248-2adb5c5e73b6 req-460a6c27-1e5e-48c7-a974-e5c082165a98 service nova] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Received event network-vif-deleted-b7baad22-e600-46ce-8fda-95751fad5c40 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.621723] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "refresh_cache-15120438-f2a8-4295-a70f-6ec8afdd9030" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.621723] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 792.621723] env[63197]: DEBUG nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.621723] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.636473] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.746949] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Releasing lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.747388] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.747590] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 792.747869] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb994b89-b5ed-43d3-95f9-08cee37d778c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.758719] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed07aa3-0037-457a-8463-90f6b14c9724 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.778118] env[63197]: DEBUG nova.compute.manager [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Received event network-changed-9a3f1efe-1c8d-4566-b375-1d15b136f104 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.778268] env[63197]: DEBUG nova.compute.manager [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Refreshing instance network info cache due to event network-changed-9a3f1efe-1c8d-4566-b375-1d15b136f104. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 792.778519] env[63197]: DEBUG oslo_concurrency.lockutils [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] Acquiring lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.778645] env[63197]: DEBUG oslo_concurrency.lockutils [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] Acquired lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.778747] env[63197]: DEBUG nova.network.neutron [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Refreshing network info cache for port 9a3f1efe-1c8d-4566-b375-1d15b136f104 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.784705] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 82919ffb-2b50-4336-9517-c741d259f19e could not be found. [ 792.784705] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 792.784705] env[63197]: INFO nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 792.784705] env[63197]: DEBUG oslo.service.loopingcall [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.786647] env[63197]: DEBUG nova.compute.manager [-] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.786744] env[63197]: DEBUG nova.network.neutron [-] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.801756] env[63197]: DEBUG nova.network.neutron [-] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.808818] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e68dc6-043c-4cda-9a53-013e57fa59c3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.817066] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3805df-53f0-4541-a4bb-2015885de8bd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.847764] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5ed57b-d06f-4963-b9c3-1c3308be516b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.855147] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d68139b-2fcc-4660-a0a9-ef401b99e43b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.868087] env[63197]: DEBUG nova.compute.provider_tree [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.140272] env[63197]: DEBUG nova.network.neutron [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.302822] env[63197]: DEBUG nova.network.neutron [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.304637] env[63197]: DEBUG nova.network.neutron [-] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.371421] env[63197]: DEBUG nova.scheduler.client.report [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.387681] env[63197]: DEBUG nova.network.neutron [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.642956] env[63197]: INFO nova.compute.manager [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 15120438-f2a8-4295-a70f-6ec8afdd9030] Took 1.02 seconds to deallocate network for instance. [ 793.809444] env[63197]: INFO nova.compute.manager [-] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Took 1.02 seconds to deallocate network for instance. [ 793.811936] env[63197]: DEBUG nova.compute.claims [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 793.812277] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.876552] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.877073] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.879568] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.680s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.891026] env[63197]: DEBUG oslo_concurrency.lockutils [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] Releasing lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.891226] env[63197]: DEBUG nova.compute.manager [req-20f241ce-dc66-4042-93ab-657169819ca3 req-2635e2d7-96da-4fb9-84fb-3cdc74fba1d9 service nova] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Received event network-vif-deleted-9a3f1efe-1c8d-4566-b375-1d15b136f104 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.384585] env[63197]: DEBUG nova.compute.utils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.388689] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.388882] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 794.440364] env[63197]: DEBUG nova.policy [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b9ac8e8a3ae240d6aec71fa495ce05d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea9e0658e1784e4c88f1794ba7cbdf0a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.669377] env[63197]: INFO nova.scheduler.client.report [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleted allocations for instance 15120438-f2a8-4295-a70f-6ec8afdd9030 [ 794.686470] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00390ba2-f7a4-4878-aa87-ab6e02ac9fb3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.697055] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d0bdbf-f767-4e28-89d5-ebfbda5db8c2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.729796] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Successfully created port: 87c0da7a-9db1-4915-8f05-eaa5ef8cc852 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.732108] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5ed6b4-3e30-4939-a8b4-ed5e522446cb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.739770] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad9db2a-dda5-4db1-bd0d-82c9adfd01a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.753081] env[63197]: DEBUG nova.compute.provider_tree [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.891370] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 795.181208] env[63197]: DEBUG oslo_concurrency.lockutils [None req-64793093-ea6b-4284-8c09-a4130925ab60 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "15120438-f2a8-4295-a70f-6ec8afdd9030" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.708s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.255944] env[63197]: DEBUG nova.scheduler.client.report [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.548841] env[63197]: DEBUG nova.compute.manager [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Received event network-changed-87c0da7a-9db1-4915-8f05-eaa5ef8cc852 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.548841] env[63197]: DEBUG nova.compute.manager [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Refreshing instance network info cache due to event network-changed-87c0da7a-9db1-4915-8f05-eaa5ef8cc852. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.549255] env[63197]: DEBUG oslo_concurrency.lockutils [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] Acquiring lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.549255] env[63197]: DEBUG oslo_concurrency.lockutils [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] Acquired lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.549334] env[63197]: DEBUG nova.network.neutron [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Refreshing network info cache for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 795.684278] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.717151] env[63197]: ERROR nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. [ 795.717151] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 795.717151] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.717151] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 795.717151] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.717151] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 795.717151] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.717151] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 795.717151] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.717151] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 795.717151] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.717151] env[63197]: ERROR nova.compute.manager raise self.value [ 795.717151] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.717151] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 795.717151] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.717151] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 795.717879] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.717879] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 795.717879] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. [ 795.717879] env[63197]: ERROR nova.compute.manager [ 795.717879] env[63197]: Traceback (most recent call last): [ 795.717879] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 795.717879] env[63197]: listener.cb(fileno) [ 795.717879] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.717879] env[63197]: result = function(*args, **kwargs) [ 795.717879] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.717879] env[63197]: return func(*args, **kwargs) [ 795.717879] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.717879] env[63197]: raise e [ 795.717879] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.717879] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 795.717879] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.717879] env[63197]: created_port_ids = self._update_ports_for_instance( [ 795.717879] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.717879] env[63197]: with excutils.save_and_reraise_exception(): [ 795.717879] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.717879] env[63197]: self.force_reraise() [ 795.717879] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.717879] env[63197]: raise self.value [ 795.717879] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.717879] env[63197]: updated_port = self._update_port( [ 795.717879] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.717879] env[63197]: _ensure_no_port_binding_failure(port) [ 795.717879] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.717879] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 795.719616] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. [ 795.719616] env[63197]: Removing descriptor: 18 [ 795.763486] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.764115] env[63197]: ERROR nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Traceback (most recent call last): [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self.driver.spawn(context, instance, image_meta, [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] vm_ref = self.build_virtual_machine(instance, [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.764115] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] for vif in network_info: [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] return self._sync_wrapper(fn, *args, **kwargs) [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self.wait() [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self[:] = self._gt.wait() [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] return self._exit_event.wait() [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] current.throw(*self._exc) [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.764509] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] result = function(*args, **kwargs) [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] return func(*args, **kwargs) [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] raise e [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] nwinfo = self.network_api.allocate_for_instance( [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] created_port_ids = self._update_ports_for_instance( [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] with excutils.save_and_reraise_exception(): [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] self.force_reraise() [ 795.764926] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] raise self.value [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] updated_port = self._update_port( [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] _ensure_no_port_binding_failure(port) [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] raise exception.PortBindingFailed(port_id=port['id']) [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] nova.exception.PortBindingFailed: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. [ 795.765355] env[63197]: ERROR nova.compute.manager [instance: b5935562-9a77-4b92-ac73-a7af496ccced] [ 795.765355] env[63197]: DEBUG nova.compute.utils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 795.766409] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.875s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.767819] env[63197]: INFO nova.compute.claims [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.770285] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Build of instance b5935562-9a77-4b92-ac73-a7af496ccced was re-scheduled: Binding failed for port 336f3ea4-949f-4784-9335-47dbd7681814, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 795.770689] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 795.770903] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.771058] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquired lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.771243] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.916611] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.946919] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.947069] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.947235] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.947416] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.947559] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.947704] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.947917] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.948131] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.949116] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.949116] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.949116] env[63197]: DEBUG nova.virt.hardware [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.949685] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412e61a0-9f6b-4bde-b733-c03ff6975829 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.958272] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25e429d-e729-4e8a-845d-85e6f03ee103 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.972924] env[63197]: ERROR nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Traceback (most recent call last): [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] yield resources [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self.driver.spawn(context, instance, image_meta, [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] vm_ref = self.build_virtual_machine(instance, [ 795.972924] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] for vif in network_info: [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] return self._sync_wrapper(fn, *args, **kwargs) [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self.wait() [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self[:] = self._gt.wait() [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] return self._exit_event.wait() [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 795.973375] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] current.throw(*self._exc) [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] result = function(*args, **kwargs) [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] return func(*args, **kwargs) [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] raise e [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] nwinfo = self.network_api.allocate_for_instance( [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] created_port_ids = self._update_ports_for_instance( [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] with excutils.save_and_reraise_exception(): [ 795.973819] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self.force_reraise() [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] raise self.value [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] updated_port = self._update_port( [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] _ensure_no_port_binding_failure(port) [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] raise exception.PortBindingFailed(port_id=port['id']) [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] nova.exception.PortBindingFailed: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. [ 795.974269] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] [ 795.974269] env[63197]: INFO nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Terminating instance [ 795.975195] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Acquiring lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.080546] env[63197]: DEBUG nova.network.neutron [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.159939] env[63197]: DEBUG nova.network.neutron [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.207078] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.289770] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.358147] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.663738] env[63197]: DEBUG oslo_concurrency.lockutils [req-46b1d5d7-2829-4465-acd9-379ff2e017d0 req-da281d15-48d2-4c57-99be-b58ddabdb48c service nova] Releasing lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.663738] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Acquired lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.663738] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.860418] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Releasing lock "refresh_cache-b5935562-9a77-4b92-ac73-a7af496ccced" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.860573] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 796.860750] env[63197]: DEBUG nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 796.860989] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.876532] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.966799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "46fd7dec-588e-430e-b51f-9b61d9b148b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.967049] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.992947] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.993202] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.052662] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7d0140-b8ad-43fc-b9c5-14342fa9932b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.061760] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721bdfbf-c349-48f1-ae57-b92c37c03860 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.090211] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5420c6bf-a513-41fb-b777-30dc2f7e8440 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.096981] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d011f028-8d02-40a2-b51c-ac03f608f133 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.109439] env[63197]: DEBUG nova.compute.provider_tree [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.191094] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.290541] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.379567] env[63197]: DEBUG nova.network.neutron [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.582364] env[63197]: DEBUG nova.compute.manager [req-66b2b6ca-3349-43ea-98ff-d59ee4b8e3c0 req-7c6ef8e0-f7e4-4876-beed-0a5829106b5c service nova] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Received event network-vif-deleted-87c0da7a-9db1-4915-8f05-eaa5ef8cc852 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.612788] env[63197]: DEBUG nova.scheduler.client.report [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.794199] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Releasing lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.794669] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 797.794849] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.795125] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-424135f8-6e44-46ea-9160-7ee8a396f00b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.804429] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c956557f-080f-4851-87bc-dffd1b833b95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.826224] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 66af0137-22d4-4670-8225-8d9b9dd2884a could not be found. [ 797.826463] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.826644] env[63197]: INFO nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 797.826888] env[63197]: DEBUG oslo.service.loopingcall [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.827120] env[63197]: DEBUG nova.compute.manager [-] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.827217] env[63197]: DEBUG nova.network.neutron [-] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.842706] env[63197]: DEBUG nova.network.neutron [-] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.882428] env[63197]: INFO nova.compute.manager [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: b5935562-9a77-4b92-ac73-a7af496ccced] Took 1.02 seconds to deallocate network for instance. [ 798.118746] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.119715] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.123189] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.440s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.345906] env[63197]: DEBUG nova.network.neutron [-] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.633693] env[63197]: DEBUG nova.compute.utils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.636229] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 798.636507] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 798.683711] env[63197]: DEBUG nova.policy [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bbcf58c100b540d7adda7aba60474d22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ef02c3c6f3f4087befff3a49fc77262', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.852223] env[63197]: INFO nova.compute.manager [-] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Took 1.02 seconds to deallocate network for instance. [ 798.854610] env[63197]: DEBUG nova.compute.claims [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 798.854706] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.919790] env[63197]: INFO nova.scheduler.client.report [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Deleted allocations for instance b5935562-9a77-4b92-ac73-a7af496ccced [ 798.972495] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Successfully created port: 1a6b1567-b092-4755-83c0-40a12149721a {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.980605] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f6d632-8f6a-414f-80d3-9275ae312f50 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.989166] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996b4dde-55c6-4611-a414-73c1bce74fa3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.021901] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e201363-1738-4e94-bc4b-f949e897c1fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.029385] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721ee728-07c9-46c5-9527-dd59b0cec241 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.043653] env[63197]: DEBUG nova.compute.provider_tree [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.141348] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.433665] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08ec5b2c-c7b6-4e0a-97b6-eb6c2fe1ca6a tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "b5935562-9a77-4b92-ac73-a7af496ccced" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.647s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.551038] env[63197]: DEBUG nova.scheduler.client.report [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.694139] env[63197]: DEBUG nova.compute.manager [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Received event network-changed-1a6b1567-b092-4755-83c0-40a12149721a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.694139] env[63197]: DEBUG nova.compute.manager [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Refreshing instance network info cache due to event network-changed-1a6b1567-b092-4755-83c0-40a12149721a. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.694139] env[63197]: DEBUG oslo_concurrency.lockutils [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] Acquiring lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.694139] env[63197]: DEBUG oslo_concurrency.lockutils [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] Acquired lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.694139] env[63197]: DEBUG nova.network.neutron [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Refreshing network info cache for port 1a6b1567-b092-4755-83c0-40a12149721a {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 799.899500] env[63197]: ERROR nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. [ 799.899500] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 799.899500] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.899500] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 799.899500] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 799.899500] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 799.899500] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 799.899500] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 799.899500] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.899500] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 799.899500] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.899500] env[63197]: ERROR nova.compute.manager raise self.value [ 799.899500] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 799.899500] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 799.899500] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.899500] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 799.900478] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.900478] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 799.900478] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. [ 799.900478] env[63197]: ERROR nova.compute.manager [ 799.900478] env[63197]: Traceback (most recent call last): [ 799.900478] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 799.900478] env[63197]: listener.cb(fileno) [ 799.900478] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 799.900478] env[63197]: result = function(*args, **kwargs) [ 799.900478] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 799.900478] env[63197]: return func(*args, **kwargs) [ 799.900478] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 799.900478] env[63197]: raise e [ 799.900478] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 799.900478] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 799.900478] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 799.900478] env[63197]: created_port_ids = self._update_ports_for_instance( [ 799.900478] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 799.900478] env[63197]: with excutils.save_and_reraise_exception(): [ 799.900478] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 799.900478] env[63197]: self.force_reraise() [ 799.900478] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 799.900478] env[63197]: raise self.value [ 799.900478] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 799.900478] env[63197]: updated_port = self._update_port( [ 799.900478] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 799.900478] env[63197]: _ensure_no_port_binding_failure(port) [ 799.900478] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 799.900478] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 799.901440] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. [ 799.901440] env[63197]: Removing descriptor: 18 [ 799.939821] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.053300] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.054089] env[63197]: ERROR nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Traceback (most recent call last): [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self.driver.spawn(context, instance, image_meta, [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] vm_ref = self.build_virtual_machine(instance, [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 800.054089] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] for vif in network_info: [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] return self._sync_wrapper(fn, *args, **kwargs) [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self.wait() [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self[:] = self._gt.wait() [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] return self._exit_event.wait() [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] current.throw(*self._exc) [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.054681] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] result = function(*args, **kwargs) [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] return func(*args, **kwargs) [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] raise e [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] nwinfo = self.network_api.allocate_for_instance( [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] created_port_ids = self._update_ports_for_instance( [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] with excutils.save_and_reraise_exception(): [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] self.force_reraise() [ 800.055252] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] raise self.value [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] updated_port = self._update_port( [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] _ensure_no_port_binding_failure(port) [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] raise exception.PortBindingFailed(port_id=port['id']) [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] nova.exception.PortBindingFailed: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. [ 800.055655] env[63197]: ERROR nova.compute.manager [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] [ 800.055655] env[63197]: DEBUG nova.compute.utils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 800.055942] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.870s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.057321] env[63197]: INFO nova.compute.claims [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.060046] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Build of instance a90e6999-9aab-4b83-98e4-2b71681d1bb5 was re-scheduled: Binding failed for port 393667ba-b110-418e-bee0-c07aed3c8a5b, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 800.060471] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 800.060688] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Acquiring lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.060832] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Acquired lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.060989] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.152812] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.181803] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.182057] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.182241] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.182429] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.182574] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.182717] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.182916] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.183083] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.183249] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.183406] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.183576] env[63197]: DEBUG nova.virt.hardware [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.184443] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01736d53-4c42-4ab2-a678-2aa9a808daff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.192740] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95844be2-308f-4953-95e6-a0eda13d129b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.211617] env[63197]: ERROR nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Traceback (most recent call last): [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] yield resources [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self.driver.spawn(context, instance, image_meta, [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] vm_ref = self.build_virtual_machine(instance, [ 800.211617] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] vif_infos = vmwarevif.get_vif_info(self._session, [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] for vif in network_info: [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] return self._sync_wrapper(fn, *args, **kwargs) [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self.wait() [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self[:] = self._gt.wait() [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] return self._exit_event.wait() [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 800.212075] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] current.throw(*self._exc) [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] result = function(*args, **kwargs) [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] return func(*args, **kwargs) [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] raise e [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] nwinfo = self.network_api.allocate_for_instance( [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] created_port_ids = self._update_ports_for_instance( [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] with excutils.save_and_reraise_exception(): [ 800.212580] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self.force_reraise() [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] raise self.value [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] updated_port = self._update_port( [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] _ensure_no_port_binding_failure(port) [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] raise exception.PortBindingFailed(port_id=port['id']) [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] nova.exception.PortBindingFailed: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. [ 800.213068] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] [ 800.213068] env[63197]: INFO nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Terminating instance [ 800.213916] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Acquiring lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.222299] env[63197]: DEBUG nova.network.neutron [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.347740] env[63197]: DEBUG nova.network.neutron [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.463508] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.586856] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.729841] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.850785] env[63197]: DEBUG oslo_concurrency.lockutils [req-5186cc38-bd82-46d3-a011-fa092cca193d req-c6a38113-8002-4601-b33e-44ec548a4987 service nova] Releasing lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.851267] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Acquired lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.851459] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.236649] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Releasing lock "refresh_cache-a90e6999-9aab-4b83-98e4-2b71681d1bb5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.237017] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 801.237017] env[63197]: DEBUG nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 801.237097] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 801.253198] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.350489] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0017406b-47df-40b1-a96f-e2291717b5d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.360017] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa05231-81ad-4fcc-9219-65e7aeac9f6a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.389968] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.392166] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf71fbed-0736-4a30-9f3b-61a4c5096e04 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.399719] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0d5cea-1998-45c7-993c-8f423984e83c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.414565] env[63197]: DEBUG nova.compute.provider_tree [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.488772] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.733586] env[63197]: DEBUG nova.compute.manager [req-c92ebf09-a536-4569-abb3-c03be1c16541 req-dfc4ce36-8401-413f-921b-29b05af5b19e service nova] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Received event network-vif-deleted-1a6b1567-b092-4755-83c0-40a12149721a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.754697] env[63197]: DEBUG nova.network.neutron [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.917919] env[63197]: DEBUG nova.scheduler.client.report [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.988226] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Releasing lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.988668] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 801.988903] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 801.989226] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3bb93767-109b-48c2-bf7d-db468baf1241 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.998099] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7960bcd6-d7a4-4b2d-89d7-b63f7eea692f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.018805] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe could not be found. [ 802.019060] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 802.019243] env[63197]: INFO nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Took 0.03 seconds to destroy the instance on the hypervisor. [ 802.019489] env[63197]: DEBUG oslo.service.loopingcall [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.019705] env[63197]: DEBUG nova.compute.manager [-] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 802.019799] env[63197]: DEBUG nova.network.neutron [-] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.032865] env[63197]: DEBUG nova.network.neutron [-] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.136744] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.136975] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.257614] env[63197]: INFO nova.compute.manager [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] [instance: a90e6999-9aab-4b83-98e4-2b71681d1bb5] Took 1.02 seconds to deallocate network for instance. [ 802.423145] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.423659] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.426368] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.326s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.427705] env[63197]: INFO nova.compute.claims [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.535748] env[63197]: DEBUG nova.network.neutron [-] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.934565] env[63197]: DEBUG nova.compute.utils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.937836] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.938035] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 802.982926] env[63197]: DEBUG nova.policy [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b50be4b75a94b4481c9c65ea1e4e9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bb89fb32d8c4726a9a3104d68ce560a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.038551] env[63197]: INFO nova.compute.manager [-] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Took 1.02 seconds to deallocate network for instance. [ 803.040996] env[63197]: DEBUG nova.compute.claims [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 803.041219] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.251859] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Successfully created port: a742b76d-91fa-4929-8d57-def111bf3bc5 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.288791] env[63197]: INFO nova.scheduler.client.report [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Deleted allocations for instance a90e6999-9aab-4b83-98e4-2b71681d1bb5 [ 803.438246] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.784290] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d2e750-3425-4463-85fb-dec7d7b60866 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.797748] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9045557e-16f1-49b0-a552-c0d86c2450cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.803043] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e528d63b-504c-4d0e-8a9e-c3aff212c75d tempest-ServerMetadataNegativeTestJSON-2075768695 tempest-ServerMetadataNegativeTestJSON-2075768695-project-member] Lock "a90e6999-9aab-4b83-98e4-2b71681d1bb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.089s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.832077] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c4bcd0-6999-449f-9f29-cc06e8ffd3ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.842061] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe81110f-9a52-4513-aca1-724422b231ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.859414] env[63197]: DEBUG nova.compute.provider_tree [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.052031] env[63197]: DEBUG nova.compute.manager [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Received event network-changed-a742b76d-91fa-4929-8d57-def111bf3bc5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.052270] env[63197]: DEBUG nova.compute.manager [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Refreshing instance network info cache due to event network-changed-a742b76d-91fa-4929-8d57-def111bf3bc5. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.052498] env[63197]: DEBUG oslo_concurrency.lockutils [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] Acquiring lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.052638] env[63197]: DEBUG oslo_concurrency.lockutils [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] Acquired lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.052792] env[63197]: DEBUG nova.network.neutron [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Refreshing network info cache for port a742b76d-91fa-4929-8d57-def111bf3bc5 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.211792] env[63197]: ERROR nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. [ 804.211792] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 804.211792] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.211792] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 804.211792] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.211792] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 804.211792] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.211792] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 804.211792] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.211792] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 804.211792] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.211792] env[63197]: ERROR nova.compute.manager raise self.value [ 804.211792] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.211792] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 804.211792] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.211792] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 804.212369] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.212369] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 804.212369] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. [ 804.212369] env[63197]: ERROR nova.compute.manager [ 804.212369] env[63197]: Traceback (most recent call last): [ 804.212369] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 804.212369] env[63197]: listener.cb(fileno) [ 804.212369] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.212369] env[63197]: result = function(*args, **kwargs) [ 804.212369] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.212369] env[63197]: return func(*args, **kwargs) [ 804.212369] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.212369] env[63197]: raise e [ 804.212369] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.212369] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 804.212369] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.212369] env[63197]: created_port_ids = self._update_ports_for_instance( [ 804.212369] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.212369] env[63197]: with excutils.save_and_reraise_exception(): [ 804.212369] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.212369] env[63197]: self.force_reraise() [ 804.212369] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.212369] env[63197]: raise self.value [ 804.212369] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.212369] env[63197]: updated_port = self._update_port( [ 804.212369] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.212369] env[63197]: _ensure_no_port_binding_failure(port) [ 804.212369] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.212369] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 804.213258] env[63197]: nova.exception.PortBindingFailed: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. [ 804.213258] env[63197]: Removing descriptor: 18 [ 804.304834] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 804.363332] env[63197]: DEBUG nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.456109] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.486780] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.487058] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.487220] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.487571] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.487858] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.488242] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.488538] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.488772] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.489477] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.489878] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.490194] env[63197]: DEBUG nova.virt.hardware [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.491452] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88433da-5616-4a35-b17d-facb986861e4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.501447] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ab49ed-2f91-417c-960f-c1594f426b79 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.517986] env[63197]: ERROR nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Traceback (most recent call last): [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] yield resources [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self.driver.spawn(context, instance, image_meta, [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] vm_ref = self.build_virtual_machine(instance, [ 804.517986] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] for vif in network_info: [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] return self._sync_wrapper(fn, *args, **kwargs) [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self.wait() [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self[:] = self._gt.wait() [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] return self._exit_event.wait() [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 804.518415] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] current.throw(*self._exc) [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] result = function(*args, **kwargs) [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] return func(*args, **kwargs) [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] raise e [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] nwinfo = self.network_api.allocate_for_instance( [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] created_port_ids = self._update_ports_for_instance( [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] with excutils.save_and_reraise_exception(): [ 804.518810] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self.force_reraise() [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] raise self.value [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] updated_port = self._update_port( [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] _ensure_no_port_binding_failure(port) [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] raise exception.PortBindingFailed(port_id=port['id']) [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] nova.exception.PortBindingFailed: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. [ 804.519205] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] [ 804.519205] env[63197]: INFO nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Terminating instance [ 804.521027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.574615] env[63197]: DEBUG nova.network.neutron [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.659380] env[63197]: DEBUG nova.network.neutron [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.826941] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.869027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.869027] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.872301] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.645s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.872497] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.872652] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 804.873325] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.888s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.877574] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468e96a6-18bb-4c5b-a299-73d27718830d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.885718] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03b6733-8a5d-4ebd-9624-04ff19d99e76 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.900145] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca40a256-0f62-4bbe-9569-c6bd0d42ee98 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.907011] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed7b279-586c-4ee3-a7b1-8b8c2c9f46bc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.937640] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181493MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 804.937828] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.161311] env[63197]: DEBUG oslo_concurrency.lockutils [req-162967d2-a605-408d-acd5-51352cb428f8 req-218bc6c3-9170-471b-b8bb-bdd7c48aaa22 service nova] Releasing lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.161732] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.161917] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.379414] env[63197]: DEBUG nova.compute.utils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.381666] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.381766] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.428714] env[63197]: DEBUG nova.policy [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8328fbf6a9004857a89bf70e99bc670a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b3ded499fcb485f9567c96500d65a23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.682615] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.687302] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca0f718-082f-43a1-bce0-555ffb1263ea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.695019] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee76a9b-0570-486d-814c-f3747e5a892e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.726574] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8594733-634e-4b69-a2d4-0e54b93a61a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.734122] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a062633-6325-45ea-bf77-a698a830ac84 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.748847] env[63197]: DEBUG nova.compute.provider_tree [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.770570] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Successfully created port: 0c4c40f4-2c3a-412d-b444-187d6badf45a {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.775020] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.887769] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.080968] env[63197]: DEBUG nova.compute.manager [req-766f3467-dd29-4e88-b232-f1b4bd9efb3d req-a276aeb1-92b4-4cdf-b74d-7f4d77133258 service nova] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Received event network-vif-deleted-a742b76d-91fa-4929-8d57-def111bf3bc5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.251667] env[63197]: DEBUG nova.scheduler.client.report [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.275464] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.275958] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 806.276631] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 806.277160] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d75eeb7c-f041-425c-bc24-dd1544cf7225 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.290051] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17cd59d-f1de-492a-b520-118475bf7e29 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.316305] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba could not be found. [ 806.316490] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.316673] env[63197]: INFO nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Took 0.04 seconds to destroy the instance on the hypervisor. [ 806.316913] env[63197]: DEBUG oslo.service.loopingcall [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.317151] env[63197]: DEBUG nova.compute.manager [-] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.317262] env[63197]: DEBUG nova.network.neutron [-] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 806.333654] env[63197]: DEBUG nova.network.neutron [-] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.688929] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. [ 806.688929] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 806.688929] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.688929] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 806.688929] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.688929] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 806.688929] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.688929] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 806.688929] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.688929] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 806.688929] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.688929] env[63197]: ERROR nova.compute.manager raise self.value [ 806.688929] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.688929] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 806.688929] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.688929] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 806.689680] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.689680] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 806.689680] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. [ 806.689680] env[63197]: ERROR nova.compute.manager [ 806.689680] env[63197]: Traceback (most recent call last): [ 806.689680] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 806.689680] env[63197]: listener.cb(fileno) [ 806.689680] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.689680] env[63197]: result = function(*args, **kwargs) [ 806.689680] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 806.689680] env[63197]: return func(*args, **kwargs) [ 806.689680] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.689680] env[63197]: raise e [ 806.689680] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.689680] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 806.689680] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.689680] env[63197]: created_port_ids = self._update_ports_for_instance( [ 806.689680] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.689680] env[63197]: with excutils.save_and_reraise_exception(): [ 806.689680] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.689680] env[63197]: self.force_reraise() [ 806.689680] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.689680] env[63197]: raise self.value [ 806.689680] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.689680] env[63197]: updated_port = self._update_port( [ 806.689680] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.689680] env[63197]: _ensure_no_port_binding_failure(port) [ 806.689680] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.689680] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 806.690660] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. [ 806.690660] env[63197]: Removing descriptor: 18 [ 806.758220] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.885s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.758857] env[63197]: ERROR nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Traceback (most recent call last): [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self.driver.spawn(context, instance, image_meta, [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] vm_ref = self.build_virtual_machine(instance, [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 806.758857] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] for vif in network_info: [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return self._sync_wrapper(fn, *args, **kwargs) [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self.wait() [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self[:] = self._gt.wait() [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return self._exit_event.wait() [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] result = hub.switch() [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 806.759295] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return self.greenlet.switch() [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] result = function(*args, **kwargs) [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] return func(*args, **kwargs) [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] raise e [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] nwinfo = self.network_api.allocate_for_instance( [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] created_port_ids = self._update_ports_for_instance( [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] with excutils.save_and_reraise_exception(): [ 806.759711] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] self.force_reraise() [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] raise self.value [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] updated_port = self._update_port( [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] _ensure_no_port_binding_failure(port) [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] raise exception.PortBindingFailed(port_id=port['id']) [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] nova.exception.PortBindingFailed: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. [ 806.760125] env[63197]: ERROR nova.compute.manager [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] [ 806.760543] env[63197]: DEBUG nova.compute.utils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 806.760861] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.508s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.763748] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Build of instance bd771a95-98b1-4fdb-b213-42d693e027f6 was re-scheduled: Binding failed for port e0ec3f0b-e3c3-457c-b087-79b0ed2f457b, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 806.764172] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 806.764397] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.764546] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquired lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.764698] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.835923] env[63197]: DEBUG nova.network.neutron [-] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.897619] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.925936] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.925936] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.925936] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.926494] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.926494] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.926494] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.926494] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.926494] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.926759] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.926759] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.926759] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.926759] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a0cfe9-c691-4421-bd50-684168a53574 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.933768] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13eb205d-eef0-4c3f-9cfe-8c84d67eab74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.948759] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Traceback (most recent call last): [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] yield resources [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self.driver.spawn(context, instance, image_meta, [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] vm_ref = self.build_virtual_machine(instance, [ 806.948759] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] vif_infos = vmwarevif.get_vif_info(self._session, [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] for vif in network_info: [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] return self._sync_wrapper(fn, *args, **kwargs) [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self.wait() [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self[:] = self._gt.wait() [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] return self._exit_event.wait() [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 806.949224] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] current.throw(*self._exc) [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] result = function(*args, **kwargs) [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] return func(*args, **kwargs) [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] raise e [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] nwinfo = self.network_api.allocate_for_instance( [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] created_port_ids = self._update_ports_for_instance( [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] with excutils.save_and_reraise_exception(): [ 806.949656] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self.force_reraise() [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] raise self.value [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] updated_port = self._update_port( [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] _ensure_no_port_binding_failure(port) [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] raise exception.PortBindingFailed(port_id=port['id']) [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] nova.exception.PortBindingFailed: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. [ 806.950022] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] [ 806.950022] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Terminating instance [ 806.951106] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.951298] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquired lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.951476] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.285704] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.338661] env[63197]: INFO nova.compute.manager [-] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Took 1.02 seconds to deallocate network for instance. [ 807.343199] env[63197]: DEBUG nova.compute.claims [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 807.343390] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.386949] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.469912] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.530808] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e14502-4fd3-4dbb-bd49-04ee4118c41b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.538948] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722514a5-8931-4543-85f4-c72934ac08da {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.567975] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.570331] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b5e779-79ba-41b9-ae14-ff7ef09001e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.577751] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4abdec8f-dce5-4b4b-8d6f-96d9f8b1f27d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.591818] env[63197]: DEBUG nova.compute.provider_tree [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.890145] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Releasing lock "refresh_cache-bd771a95-98b1-4fdb-b213-42d693e027f6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.890361] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 807.890361] env[63197]: DEBUG nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 807.890515] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 807.906459] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.071025] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Releasing lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.071412] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 808.071719] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 808.072154] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdb7ca93-186c-4d33-8324-557d744aefc2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.082859] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab91b4b-d23e-4ebf-9cfb-590664b13969 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.095535] env[63197]: DEBUG nova.scheduler.client.report [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.102711] env[63197]: DEBUG nova.compute.manager [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Received event network-changed-0c4c40f4-2c3a-412d-b444-187d6badf45a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.102891] env[63197]: DEBUG nova.compute.manager [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Refreshing instance network info cache due to event network-changed-0c4c40f4-2c3a-412d-b444-187d6badf45a. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.103345] env[63197]: DEBUG oslo_concurrency.lockutils [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] Acquiring lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.103345] env[63197]: DEBUG oslo_concurrency.lockutils [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] Acquired lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.103527] env[63197]: DEBUG nova.network.neutron [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Refreshing network info cache for port 0c4c40f4-2c3a-412d-b444-187d6badf45a {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 808.109791] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7786e5f4-a7c9-4c14-b407-d2eda85bf37a could not be found. [ 808.109989] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.110160] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 808.110388] env[63197]: DEBUG oslo.service.loopingcall [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.111265] env[63197]: DEBUG nova.compute.manager [-] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.111265] env[63197]: DEBUG nova.network.neutron [-] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.126464] env[63197]: DEBUG nova.network.neutron [-] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.409572] env[63197]: DEBUG nova.network.neutron [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.600627] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.840s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.601354] env[63197]: ERROR nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Traceback (most recent call last): [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self.driver.spawn(context, instance, image_meta, [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] vm_ref = self.build_virtual_machine(instance, [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 808.601354] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] for vif in network_info: [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return self._sync_wrapper(fn, *args, **kwargs) [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self.wait() [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self[:] = self._gt.wait() [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return self._exit_event.wait() [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] result = hub.switch() [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 808.601760] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return self.greenlet.switch() [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] result = function(*args, **kwargs) [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] return func(*args, **kwargs) [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] raise e [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] nwinfo = self.network_api.allocate_for_instance( [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] created_port_ids = self._update_ports_for_instance( [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] with excutils.save_and_reraise_exception(): [ 808.602361] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] self.force_reraise() [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] raise self.value [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] updated_port = self._update_port( [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] _ensure_no_port_binding_failure(port) [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] raise exception.PortBindingFailed(port_id=port['id']) [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] nova.exception.PortBindingFailed: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. [ 808.602856] env[63197]: ERROR nova.compute.manager [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] [ 808.603231] env[63197]: DEBUG nova.compute.utils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 808.603293] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.791s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.606415] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Build of instance d7a2de2b-d14b-437c-93ff-17ee24bd97ee was re-scheduled: Binding failed for port b7baad22-e600-46ce-8fda-95751fad5c40, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 808.606874] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 808.607111] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.607258] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquired lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.607414] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.623589] env[63197]: DEBUG nova.network.neutron [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.628857] env[63197]: DEBUG nova.network.neutron [-] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.701409] env[63197]: DEBUG nova.network.neutron [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.914452] env[63197]: INFO nova.compute.manager [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: bd771a95-98b1-4fdb-b213-42d693e027f6] Took 1.02 seconds to deallocate network for instance. [ 809.128987] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.130507] env[63197]: INFO nova.compute.manager [-] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Took 1.02 seconds to deallocate network for instance. [ 809.134842] env[63197]: DEBUG nova.compute.claims [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 809.135035] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.204272] env[63197]: DEBUG oslo_concurrency.lockutils [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] Releasing lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.204524] env[63197]: DEBUG nova.compute.manager [req-e4dca2d3-9f0f-4d58-b4e9-005e689031e6 req-8de3473e-c640-44be-917b-89873ad6fa37 service nova] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Received event network-vif-deleted-0c4c40f4-2c3a-412d-b444-187d6badf45a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.215261] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.367535] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492a1083-4216-4880-9a24-0a0d80ef3a54 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.375274] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f582895-8245-4a89-a246-2bba76f1d477 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.404620] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8fa595a-a48a-430e-91ef-ec7e50a8d8a2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.411406] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0353cfd7-98c0-40f1-b844-e342026ddf9e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.426979] env[63197]: DEBUG nova.compute.provider_tree [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.719602] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Releasing lock "refresh_cache-d7a2de2b-d14b-437c-93ff-17ee24bd97ee" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.719602] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 809.719602] env[63197]: DEBUG nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 809.719602] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.735204] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.933358] env[63197]: DEBUG nova.scheduler.client.report [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.950185] env[63197]: INFO nova.scheduler.client.report [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Deleted allocations for instance bd771a95-98b1-4fdb-b213-42d693e027f6 [ 810.237586] env[63197]: DEBUG nova.network.neutron [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.437841] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.834s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.438490] env[63197]: ERROR nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Traceback (most recent call last): [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self.driver.spawn(context, instance, image_meta, [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] vm_ref = self.build_virtual_machine(instance, [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] vif_infos = vmwarevif.get_vif_info(self._session, [ 810.438490] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] for vif in network_info: [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] return self._sync_wrapper(fn, *args, **kwargs) [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self.wait() [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self[:] = self._gt.wait() [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] return self._exit_event.wait() [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] current.throw(*self._exc) [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 810.438889] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] result = function(*args, **kwargs) [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] return func(*args, **kwargs) [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] raise e [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] nwinfo = self.network_api.allocate_for_instance( [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] created_port_ids = self._update_ports_for_instance( [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] with excutils.save_and_reraise_exception(): [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] self.force_reraise() [ 810.439381] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] raise self.value [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] updated_port = self._update_port( [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] _ensure_no_port_binding_failure(port) [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] raise exception.PortBindingFailed(port_id=port['id']) [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] nova.exception.PortBindingFailed: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. [ 810.439801] env[63197]: ERROR nova.compute.manager [instance: 82919ffb-2b50-4336-9517-c741d259f19e] [ 810.439801] env[63197]: DEBUG nova.compute.utils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 810.440483] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.234s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.442035] env[63197]: INFO nova.compute.claims [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.444637] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Build of instance 82919ffb-2b50-4336-9517-c741d259f19e was re-scheduled: Binding failed for port 9a3f1efe-1c8d-4566-b375-1d15b136f104, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 810.445058] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 810.445284] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquiring lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.445476] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Acquired lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.445593] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 810.460410] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ee0f17d5-f1b1-47e1-a1ec-7b6c78b5e40e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "bd771a95-98b1-4fdb-b213-42d693e027f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.240s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.740653] env[63197]: INFO nova.compute.manager [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: d7a2de2b-d14b-437c-93ff-17ee24bd97ee] Took 1.02 seconds to deallocate network for instance. [ 810.962905] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.966462] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.047325] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.486371] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.550348] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Releasing lock "refresh_cache-82919ffb-2b50-4336-9517-c741d259f19e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.551263] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 811.551263] env[63197]: DEBUG nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 811.551263] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 811.567358] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.698335] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f29f484-90a4-4afb-b423-0988085545ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.706242] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24441798-360e-4cdb-b132-0d58fc003989 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.734673] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ac7f06-3cde-4c21-b075-886efcc9044e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.741086] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38103ea6-3af1-43b0-831d-24f56c808ffd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.755546] env[63197]: DEBUG nova.compute.provider_tree [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.769608] env[63197]: INFO nova.scheduler.client.report [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Deleted allocations for instance d7a2de2b-d14b-437c-93ff-17ee24bd97ee [ 812.070874] env[63197]: DEBUG nova.network.neutron [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.258940] env[63197]: DEBUG nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.278901] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b399c7d4-21a8-4a40-81f0-8884c09d3633 tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "d7a2de2b-d14b-437c-93ff-17ee24bd97ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.497s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.573130] env[63197]: INFO nova.compute.manager [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] [instance: 82919ffb-2b50-4336-9517-c741d259f19e] Took 1.02 seconds to deallocate network for instance. [ 812.764092] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.764682] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.767548] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.913s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.780642] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.272048] env[63197]: DEBUG nova.compute.utils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.275930] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 813.276108] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.297701] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.316020] env[63197]: DEBUG nova.policy [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8328fbf6a9004857a89bf70e99bc670a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b3ded499fcb485f9567c96500d65a23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.535166] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3498b5d-a32d-4aab-a30c-a99eb7454212 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.541700] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b447bb9-eb84-4ad3-8e7d-03e978e2bcf7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.570812] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f487a9ec-2f28-4eae-95c0-aaebdfb953d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.577659] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90c8d6c-af2c-4034-b107-6f2fdaf30f08 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.593744] env[63197]: DEBUG nova.compute.provider_tree [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.600493] env[63197]: INFO nova.scheduler.client.report [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Deleted allocations for instance 82919ffb-2b50-4336-9517-c741d259f19e [ 813.630718] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Successfully created port: 4bd7fc22-273e-4c8b-a00d-ec7a88087c22 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.776568] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.097843] env[63197]: DEBUG nova.scheduler.client.report [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.108809] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e868fdd9-e372-4b89-9eb4-448f1280da1e tempest-ListServerFiltersTestJSON-268123202 tempest-ListServerFiltersTestJSON-268123202-project-member] Lock "82919ffb-2b50-4336-9517-c741d259f19e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.550s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.324791] env[63197]: DEBUG nova.compute.manager [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Received event network-changed-4bd7fc22-273e-4c8b-a00d-ec7a88087c22 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.325051] env[63197]: DEBUG nova.compute.manager [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Refreshing instance network info cache due to event network-changed-4bd7fc22-273e-4c8b-a00d-ec7a88087c22. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.325218] env[63197]: DEBUG oslo_concurrency.lockutils [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] Acquiring lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.325361] env[63197]: DEBUG oslo_concurrency.lockutils [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] Acquired lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.325517] env[63197]: DEBUG nova.network.neutron [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Refreshing network info cache for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 814.517703] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. [ 814.517703] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 814.517703] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.517703] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 814.517703] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.517703] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 814.517703] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.517703] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 814.517703] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.517703] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 814.517703] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.517703] env[63197]: ERROR nova.compute.manager raise self.value [ 814.517703] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.517703] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 814.517703] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.517703] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 814.518286] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.518286] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 814.518286] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. [ 814.518286] env[63197]: ERROR nova.compute.manager [ 814.518286] env[63197]: Traceback (most recent call last): [ 814.518286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 814.518286] env[63197]: listener.cb(fileno) [ 814.518286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.518286] env[63197]: result = function(*args, **kwargs) [ 814.518286] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.518286] env[63197]: return func(*args, **kwargs) [ 814.518286] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.518286] env[63197]: raise e [ 814.518286] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.518286] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 814.518286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.518286] env[63197]: created_port_ids = self._update_ports_for_instance( [ 814.518286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.518286] env[63197]: with excutils.save_and_reraise_exception(): [ 814.518286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.518286] env[63197]: self.force_reraise() [ 814.518286] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.518286] env[63197]: raise self.value [ 814.518286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.518286] env[63197]: updated_port = self._update_port( [ 814.518286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.518286] env[63197]: _ensure_no_port_binding_failure(port) [ 814.518286] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.518286] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 814.519267] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. [ 814.519267] env[63197]: Removing descriptor: 18 [ 814.605083] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.837s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.605986] env[63197]: ERROR nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Traceback (most recent call last): [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self.driver.spawn(context, instance, image_meta, [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] vm_ref = self.build_virtual_machine(instance, [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.605986] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] for vif in network_info: [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] return self._sync_wrapper(fn, *args, **kwargs) [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self.wait() [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self[:] = self._gt.wait() [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] return self._exit_event.wait() [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] current.throw(*self._exc) [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.606440] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] result = function(*args, **kwargs) [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] return func(*args, **kwargs) [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] raise e [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] nwinfo = self.network_api.allocate_for_instance( [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] created_port_ids = self._update_ports_for_instance( [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] with excutils.save_and_reraise_exception(): [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] self.force_reraise() [ 814.606857] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] raise self.value [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] updated_port = self._update_port( [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] _ensure_no_port_binding_failure(port) [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] raise exception.PortBindingFailed(port_id=port['id']) [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] nova.exception.PortBindingFailed: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. [ 814.607281] env[63197]: ERROR nova.compute.manager [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] [ 814.607281] env[63197]: DEBUG nova.compute.utils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.608698] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.145s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.609626] env[63197]: INFO nova.compute.claims [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.616028] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Build of instance 66af0137-22d4-4670-8225-8d9b9dd2884a was re-scheduled: Binding failed for port 87c0da7a-9db1-4915-8f05-eaa5ef8cc852, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 814.616028] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 814.616028] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Acquiring lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.616028] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Acquired lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.616266] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.616266] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 814.785857] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 814.811481] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.811754] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.811907] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.812106] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.812267] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.812444] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.812651] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.812809] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.812999] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.813189] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.813458] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.814391] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639f55df-cd29-460c-bb11-77681eb47851 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.824026] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0894b00c-6dcb-4e4b-a3c9-b527918456ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.841029] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Traceback (most recent call last): [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] yield resources [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self.driver.spawn(context, instance, image_meta, [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] vm_ref = self.build_virtual_machine(instance, [ 814.841029] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] for vif in network_info: [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] return self._sync_wrapper(fn, *args, **kwargs) [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self.wait() [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self[:] = self._gt.wait() [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] return self._exit_event.wait() [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.841624] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] current.throw(*self._exc) [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] result = function(*args, **kwargs) [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] return func(*args, **kwargs) [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] raise e [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] nwinfo = self.network_api.allocate_for_instance( [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] created_port_ids = self._update_ports_for_instance( [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] with excutils.save_and_reraise_exception(): [ 814.841978] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self.force_reraise() [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] raise self.value [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] updated_port = self._update_port( [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] _ensure_no_port_binding_failure(port) [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] raise exception.PortBindingFailed(port_id=port['id']) [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] nova.exception.PortBindingFailed: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. [ 814.842827] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] [ 814.842827] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Terminating instance [ 814.843418] env[63197]: DEBUG nova.network.neutron [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.845638] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.928624] env[63197]: DEBUG nova.network.neutron [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.138974] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.142200] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.390915] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.431882] env[63197]: DEBUG oslo_concurrency.lockutils [req-3a8af7fd-caa1-476f-8ec3-bfe70ae8b9a7 req-e86bff63-3dfd-4626-b52c-0df0285cef93 service nova] Releasing lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.432038] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquired lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.432220] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.857936] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9309a247-8f11-4800-82e2-2a942aa87b02 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.865375] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be85977-2162-405b-a3ed-5fc9935e1907 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.893989] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Releasing lock "refresh_cache-66af0137-22d4-4670-8225-8d9b9dd2884a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.894252] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 815.894436] env[63197]: DEBUG nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.894601] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.896865] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320a3154-b1c4-417a-aba8-cba87970911f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.946782] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e415c30e-c2bc-4a87-9013-dd5ef6df6c85 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.946782] env[63197]: DEBUG nova.compute.provider_tree [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.946782] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.953139] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.078641] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.354645] env[63197]: DEBUG nova.compute.manager [req-04257f73-079c-45a9-a9cc-0bcab5c86b1b req-04434101-0f9f-4b3e-99ba-f653f9257b7d service nova] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Received event network-vif-deleted-4bd7fc22-273e-4c8b-a00d-ec7a88087c22 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.421934] env[63197]: DEBUG nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.425575] env[63197]: DEBUG nova.network.neutron [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.581319] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Releasing lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.581779] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.581991] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.582298] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d65528d0-7f83-4357-9d08-3bc444ec0723 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.592807] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b71f5b7-3db7-41e8-bbda-28d8acfabd8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.613719] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5128f0c1-da4a-4eb4-9dde-23f830800512 could not be found. [ 816.613953] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.614147] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Took 0.03 seconds to destroy the instance on the hypervisor. [ 816.614379] env[63197]: DEBUG oslo.service.loopingcall [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.614587] env[63197]: DEBUG nova.compute.manager [-] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.614685] env[63197]: DEBUG nova.network.neutron [-] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.636267] env[63197]: DEBUG nova.network.neutron [-] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.930069] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.930069] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.931324] env[63197]: INFO nova.compute.manager [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] [instance: 66af0137-22d4-4670-8225-8d9b9dd2884a] Took 1.04 seconds to deallocate network for instance. [ 816.933967] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.893s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.140335] env[63197]: DEBUG nova.network.neutron [-] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.435446] env[63197]: DEBUG nova.compute.utils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.435446] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.435446] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.519803] env[63197]: DEBUG nova.policy [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8328fbf6a9004857a89bf70e99bc670a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b3ded499fcb485f9567c96500d65a23', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.642646] env[63197]: INFO nova.compute.manager [-] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Took 1.03 seconds to deallocate network for instance. [ 817.645359] env[63197]: DEBUG nova.compute.claims [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 817.645534] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.713631] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80563795-b012-42cc-8ec9-8c26af9fe8ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.722599] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0dde23-18dc-400f-97eb-271184ff24fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.766079] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35a246e-70b2-4d88-a48c-9a4537796682 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.774445] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8045be39-4d3d-46a0-a23e-889ecaa28f5a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.788556] env[63197]: DEBUG nova.compute.provider_tree [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.942588] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.963781] env[63197]: INFO nova.scheduler.client.report [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Deleted allocations for instance 66af0137-22d4-4670-8225-8d9b9dd2884a [ 817.972785] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Successfully created port: 7475c740-8ca9-47dd-a390-89fd8189b300 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.291828] env[63197]: DEBUG nova.scheduler.client.report [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.474054] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c14ace7e-3657-4785-bcba-fa6de0fcbd70 tempest-InstanceActionsTestJSON-1559390553 tempest-InstanceActionsTestJSON-1559390553-project-member] Lock "66af0137-22d4-4670-8225-8d9b9dd2884a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.039s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.799862] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.865s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.799862] env[63197]: ERROR nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. [ 818.799862] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Traceback (most recent call last): [ 818.799862] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 818.799862] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self.driver.spawn(context, instance, image_meta, [ 818.799862] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 818.799862] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 818.799862] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 818.799862] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] vm_ref = self.build_virtual_machine(instance, [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] vif_infos = vmwarevif.get_vif_info(self._session, [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] for vif in network_info: [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] return self._sync_wrapper(fn, *args, **kwargs) [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self.wait() [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self[:] = self._gt.wait() [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] return self._exit_event.wait() [ 818.800155] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] current.throw(*self._exc) [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] result = function(*args, **kwargs) [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] return func(*args, **kwargs) [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] raise e [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] nwinfo = self.network_api.allocate_for_instance( [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] created_port_ids = self._update_ports_for_instance( [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.800419] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] with excutils.save_and_reraise_exception(): [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] self.force_reraise() [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] raise self.value [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] updated_port = self._update_port( [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] _ensure_no_port_binding_failure(port) [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] raise exception.PortBindingFailed(port_id=port['id']) [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] nova.exception.PortBindingFailed: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. [ 818.800665] env[63197]: ERROR nova.compute.manager [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] [ 818.801929] env[63197]: DEBUG nova.compute.utils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 818.802983] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.976s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.804616] env[63197]: INFO nova.compute.claims [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.807126] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Build of instance 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe was re-scheduled: Binding failed for port 1a6b1567-b092-4755-83c0-40a12149721a, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 818.807556] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 818.807781] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Acquiring lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.807927] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Acquired lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.808095] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.955262] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.977929] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 818.989264] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.989702] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.989702] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.990296] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.990296] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.990296] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.990453] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.990600] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.990759] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.990898] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.991077] env[63197]: DEBUG nova.virt.hardware [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.993915] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46861fd6-fb69-4a17-b45d-a44098c0ddcf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.003550] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218ffb52-da2e-4f0d-93e6-ba5329239514 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.093564] env[63197]: DEBUG nova.compute.manager [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Received event network-changed-7475c740-8ca9-47dd-a390-89fd8189b300 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.094897] env[63197]: DEBUG nova.compute.manager [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Refreshing instance network info cache due to event network-changed-7475c740-8ca9-47dd-a390-89fd8189b300. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.094897] env[63197]: DEBUG oslo_concurrency.lockutils [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] Acquiring lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.094897] env[63197]: DEBUG oslo_concurrency.lockutils [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] Acquired lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.094897] env[63197]: DEBUG nova.network.neutron [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Refreshing network info cache for port 7475c740-8ca9-47dd-a390-89fd8189b300 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.181773] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. [ 819.181773] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 819.181773] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.181773] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 819.181773] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.181773] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 819.181773] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.181773] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 819.181773] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.181773] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 819.181773] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.181773] env[63197]: ERROR nova.compute.manager raise self.value [ 819.181773] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.181773] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 819.181773] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.181773] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 819.182296] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.182296] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 819.182296] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. [ 819.182296] env[63197]: ERROR nova.compute.manager [ 819.182296] env[63197]: Traceback (most recent call last): [ 819.182296] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 819.182296] env[63197]: listener.cb(fileno) [ 819.182296] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.182296] env[63197]: result = function(*args, **kwargs) [ 819.182296] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.182296] env[63197]: return func(*args, **kwargs) [ 819.182296] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.182296] env[63197]: raise e [ 819.182296] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.182296] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 819.182296] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.182296] env[63197]: created_port_ids = self._update_ports_for_instance( [ 819.182296] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.182296] env[63197]: with excutils.save_and_reraise_exception(): [ 819.182296] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.182296] env[63197]: self.force_reraise() [ 819.182296] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.182296] env[63197]: raise self.value [ 819.182296] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.182296] env[63197]: updated_port = self._update_port( [ 819.182296] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.182296] env[63197]: _ensure_no_port_binding_failure(port) [ 819.182296] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.182296] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 819.182982] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. [ 819.182982] env[63197]: Removing descriptor: 18 [ 819.182982] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Traceback (most recent call last): [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] yield resources [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self.driver.spawn(context, instance, image_meta, [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 819.182982] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] vm_ref = self.build_virtual_machine(instance, [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] for vif in network_info: [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return self._sync_wrapper(fn, *args, **kwargs) [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self.wait() [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self[:] = self._gt.wait() [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return self._exit_event.wait() [ 819.183270] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] result = hub.switch() [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return self.greenlet.switch() [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] result = function(*args, **kwargs) [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return func(*args, **kwargs) [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] raise e [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] nwinfo = self.network_api.allocate_for_instance( [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 819.183561] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] created_port_ids = self._update_ports_for_instance( [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] with excutils.save_and_reraise_exception(): [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self.force_reraise() [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] raise self.value [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] updated_port = self._update_port( [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] _ensure_no_port_binding_failure(port) [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 819.183888] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] raise exception.PortBindingFailed(port_id=port['id']) [ 819.184171] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] nova.exception.PortBindingFailed: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. [ 819.184171] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] [ 819.184171] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Terminating instance [ 819.184825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.332902] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.394933] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.395201] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.425929] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.577029] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.615455] env[63197]: DEBUG nova.network.neutron [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.710981] env[63197]: DEBUG nova.network.neutron [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.835741] env[63197]: DEBUG nova.scheduler.client.report [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Refreshing inventories for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 819.860297] env[63197]: DEBUG nova.scheduler.client.report [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Updating ProviderTree inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 819.862917] env[63197]: DEBUG nova.compute.provider_tree [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 819.873854] env[63197]: DEBUG nova.scheduler.client.report [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Refreshing aggregate associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, aggregates: None {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 819.898989] env[63197]: DEBUG nova.scheduler.client.report [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Refreshing trait associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64 {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 819.927463] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Releasing lock "refresh_cache-0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.928600] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 819.928903] env[63197]: DEBUG nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.929122] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 819.946142] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.196595] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f93dfb2-f032-4c8e-a254-0d0fcb9c1482 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.206041] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0363c6-ef6d-4915-9204-aa9e49508ca1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.234433] env[63197]: DEBUG oslo_concurrency.lockutils [req-78912d1c-4171-4a76-a622-78bdac7eea0d req-137acea7-b637-48dc-aaea-446bcaad614a service nova] Releasing lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.235420] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquired lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.235611] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.237111] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7f9cc8-8a7d-4a30-b2c7-9a040553e75d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.244740] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd1d6b8-0a2e-40f0-84a8-721df727b049 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.259812] env[63197]: DEBUG nova.compute.provider_tree [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.451841] env[63197]: DEBUG nova.network.neutron [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.760114] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.762569] env[63197]: DEBUG nova.scheduler.client.report [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.853554] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.955086] env[63197]: INFO nova.compute.manager [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] [instance: 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe] Took 1.03 seconds to deallocate network for instance. [ 821.129675] env[63197]: DEBUG nova.compute.manager [req-7a3313e5-a01b-47e4-8378-a5fe5be9fc32 req-d8c3e3a3-ac36-4459-a381-a7916949111a service nova] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Received event network-vif-deleted-7475c740-8ca9-47dd-a390-89fd8189b300 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.268555] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.269612] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.332s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.356011] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Releasing lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.356460] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.356649] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.356945] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa3daeb6-bd34-41ce-af1a-db8b542bec68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.366172] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332fc282-c01a-4599-8ec1-d13f0e925059 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.386338] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06167654-0ac8-4b04-8eb7-071eac0894ae could not be found. [ 821.386540] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 821.386726] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Took 0.03 seconds to destroy the instance on the hypervisor. [ 821.386947] env[63197]: DEBUG oslo.service.loopingcall [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.387444] env[63197]: DEBUG nova.compute.manager [-] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.387537] env[63197]: DEBUG nova.network.neutron [-] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 821.401745] env[63197]: DEBUG nova.network.neutron [-] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.776021] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquiring lock "05c0fbd0-4454-4aa8-9011-dbee262bc222" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.776021] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "05c0fbd0-4454-4aa8-9011-dbee262bc222" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.904088] env[63197]: DEBUG nova.network.neutron [-] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.991663] env[63197]: INFO nova.scheduler.client.report [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Deleted allocations for instance 0ba9f984-0cf0-489c-91bc-9d3df1d30ebe [ 822.281332] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "05c0fbd0-4454-4aa8-9011-dbee262bc222" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.281332] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.301554] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 822.301709] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 7786e5f4-a7c9-4c14-b407-d2eda85bf37a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 822.301827] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 5128f0c1-da4a-4eb4-9dde-23f830800512 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 822.301941] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 06167654-0ac8-4b04-8eb7-071eac0894ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 822.302101] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance f8a0c645-4595-41e7-b564-2f74fed9f275 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 822.406369] env[63197]: INFO nova.compute.manager [-] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Took 1.02 seconds to deallocate network for instance. [ 822.408813] env[63197]: DEBUG nova.compute.claims [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 822.409053] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.502245] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f45ebb5a-4ed2-4225-a319-520af2eb22c3 tempest-ServersNegativeTestJSON-400726605 tempest-ServersNegativeTestJSON-400726605-project-member] Lock "0ba9f984-0cf0-489c-91bc-9d3df1d30ebe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.050s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.786812] env[63197]: DEBUG nova.compute.utils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.788170] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.792089] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.804807] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance fc4d4223-b0cb-4a40-bd5f-62ec740b57d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 822.839733] env[63197]: DEBUG nova.policy [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2fd057687d214cf8bac10509727c6b88', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c7c736d3d8d4cad98789a798307ea54', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.004803] env[63197]: DEBUG nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 823.143028] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Successfully created port: 4570b946-20eb-46f8-ad69-d9cfc02035fa {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.294201] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.306880] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance c005d088-ee27-4d20-9db4-65b47227f22a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 823.540567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.809870] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 643622a1-cb00-4829-a83a-a347e6e50e5d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 823.887076] env[63197]: DEBUG nova.compute.manager [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Received event network-changed-4570b946-20eb-46f8-ad69-d9cfc02035fa {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.887076] env[63197]: DEBUG nova.compute.manager [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Refreshing instance network info cache due to event network-changed-4570b946-20eb-46f8-ad69-d9cfc02035fa. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.887076] env[63197]: DEBUG oslo_concurrency.lockutils [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] Acquiring lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.887511] env[63197]: DEBUG oslo_concurrency.lockutils [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] Acquired lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.887756] env[63197]: DEBUG nova.network.neutron [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Refreshing network info cache for port 4570b946-20eb-46f8-ad69-d9cfc02035fa {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.064401] env[63197]: ERROR nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. [ 824.064401] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 824.064401] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.064401] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 824.064401] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.064401] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 824.064401] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.064401] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 824.064401] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.064401] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 824.064401] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.064401] env[63197]: ERROR nova.compute.manager raise self.value [ 824.064401] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.064401] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 824.064401] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.064401] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 824.064810] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.064810] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 824.064810] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. [ 824.064810] env[63197]: ERROR nova.compute.manager [ 824.064810] env[63197]: Traceback (most recent call last): [ 824.064810] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 824.064810] env[63197]: listener.cb(fileno) [ 824.064810] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.064810] env[63197]: result = function(*args, **kwargs) [ 824.064810] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.064810] env[63197]: return func(*args, **kwargs) [ 824.064810] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.064810] env[63197]: raise e [ 824.064810] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.064810] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 824.064810] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.064810] env[63197]: created_port_ids = self._update_ports_for_instance( [ 824.064810] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.064810] env[63197]: with excutils.save_and_reraise_exception(): [ 824.064810] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.064810] env[63197]: self.force_reraise() [ 824.064810] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.064810] env[63197]: raise self.value [ 824.064810] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.064810] env[63197]: updated_port = self._update_port( [ 824.064810] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.064810] env[63197]: _ensure_no_port_binding_failure(port) [ 824.064810] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.064810] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 824.065542] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. [ 824.065542] env[63197]: Removing descriptor: 18 [ 824.301682] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.312705] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 92ee8e7a-c169-44d8-8f74-84590b4f62e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 824.326028] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.326028] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.326189] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.327028] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.327028] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.327028] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.327028] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.327028] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.327295] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.327295] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.327454] env[63197]: DEBUG nova.virt.hardware [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.328545] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1e0c7c-49b5-405e-8f7f-a77638a893c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.337485] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ce4616-55a6-42ab-bde5-66d00e108c6a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.351890] env[63197]: ERROR nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Traceback (most recent call last): [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] yield resources [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self.driver.spawn(context, instance, image_meta, [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self._vmops.spawn(context, instance, image_meta, injected_files, [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] vm_ref = self.build_virtual_machine(instance, [ 824.351890] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] vif_infos = vmwarevif.get_vif_info(self._session, [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] for vif in network_info: [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] return self._sync_wrapper(fn, *args, **kwargs) [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self.wait() [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self[:] = self._gt.wait() [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] return self._exit_event.wait() [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 824.352326] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] current.throw(*self._exc) [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] result = function(*args, **kwargs) [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] return func(*args, **kwargs) [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] raise e [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] nwinfo = self.network_api.allocate_for_instance( [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] created_port_ids = self._update_ports_for_instance( [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] with excutils.save_and_reraise_exception(): [ 824.352731] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self.force_reraise() [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] raise self.value [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] updated_port = self._update_port( [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] _ensure_no_port_binding_failure(port) [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] raise exception.PortBindingFailed(port_id=port['id']) [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] nova.exception.PortBindingFailed: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. [ 824.353065] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] [ 824.353065] env[63197]: INFO nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Terminating instance [ 824.357773] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquiring lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.404848] env[63197]: DEBUG nova.network.neutron [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.489247] env[63197]: DEBUG nova.network.neutron [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.815539] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 65df32fa-a6e7-4c3a-af8b-0422df8d9229 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 824.992056] env[63197]: DEBUG oslo_concurrency.lockutils [req-1beae09e-3d4d-433f-86df-c1e97b496c40 req-e7ebb02b-f3ba-44c7-9e39-cc6f1d391d97 service nova] Releasing lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.992056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquired lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.992199] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.319029] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance c55b5b2b-5ecc-43bb-a279-7370cd9ac722 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 825.518175] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.640032] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.822934] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 825.911733] env[63197]: DEBUG nova.compute.manager [req-df281244-116a-406c-85ef-f7bdd8b723e2 req-628b6602-fd0b-42e8-bf9a-e9c410a4c5b4 service nova] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Received event network-vif-deleted-4570b946-20eb-46f8-ad69-d9cfc02035fa {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 826.144496] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Releasing lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.145033] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 826.145243] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 826.145541] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4184017c-d7d2-4341-aa89-6bb803c993c9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.154347] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add4c499-fd23-47be-bd87-659cc43a5895 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.174256] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f8a0c645-4595-41e7-b564-2f74fed9f275 could not be found. [ 826.174451] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 826.174624] env[63197]: INFO nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Took 0.03 seconds to destroy the instance on the hypervisor. [ 826.174847] env[63197]: DEBUG oslo.service.loopingcall [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.175054] env[63197]: DEBUG nova.compute.manager [-] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.175151] env[63197]: DEBUG nova.network.neutron [-] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 826.190643] env[63197]: DEBUG nova.network.neutron [-] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.325813] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 9c1caa2b-c369-425a-8726-cddadf06f338 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 826.693999] env[63197]: DEBUG nova.network.neutron [-] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.828578] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 72f9b0c4-69b9-49f2-8665-ff77151883af has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 827.197012] env[63197]: INFO nova.compute.manager [-] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Took 1.02 seconds to deallocate network for instance. [ 827.199842] env[63197]: DEBUG nova.compute.claims [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 827.200033] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.333348] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a0a40337-d8d7-448b-afff-f6849e9d37a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 827.836541] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 46fd7dec-588e-430e-b51f-9b61d9b148b6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 828.343365] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 8500a7e9-3fdc-411e-a48e-189d4d7bffba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 828.846462] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 829.349515] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 829.349770] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 829.349914] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 829.550298] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6612c86a-ba3c-4140-8f23-515d440ed153 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.557613] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bf7b9c4-f660-4e4f-b3a4-ba3fbdf671c6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.585719] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cd6093-63ef-4fd8-86a4-a0a259d5a6b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.592193] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36eae992-6847-4b4b-90dc-9af5df8c9254 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.604641] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.107229] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.613410] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 830.613649] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 9.344s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.613956] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.271s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.616957] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 830.617118] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Cleaning up deleted instances {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 831.124067] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] There are 4 instances to clean {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 831.124322] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 09dc9549-ab9a-4f36-9671-ab1b5595649d] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 831.321897] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b280df-505e-4ab6-bbd2-f97e0e7a126b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.329388] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90df3080-ae7d-49b1-b197-ec467197dc4d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.359856] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee9da74-5275-4d07-aced-92034ace8d40 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.365996] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4225d59f-e41f-41e9-a5df-008f2003c805 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.378941] env[63197]: DEBUG nova.compute.provider_tree [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.627587] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a066e6e1-12a4-455b-87cc-e3d6a6a17c1f] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 831.881897] env[63197]: DEBUG nova.scheduler.client.report [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.131092] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 99539d7f-cd93-4250-97cd-2449e03e517d] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 832.386981] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.773s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.387625] env[63197]: ERROR nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Traceback (most recent call last): [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self.driver.spawn(context, instance, image_meta, [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] vm_ref = self.build_virtual_machine(instance, [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 832.387625] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] for vif in network_info: [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] return self._sync_wrapper(fn, *args, **kwargs) [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self.wait() [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self[:] = self._gt.wait() [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] return self._exit_event.wait() [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] current.throw(*self._exc) [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 832.387901] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] result = function(*args, **kwargs) [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] return func(*args, **kwargs) [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] raise e [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] nwinfo = self.network_api.allocate_for_instance( [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] created_port_ids = self._update_ports_for_instance( [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] with excutils.save_and_reraise_exception(): [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] self.force_reraise() [ 832.388180] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] raise self.value [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] updated_port = self._update_port( [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] _ensure_no_port_binding_failure(port) [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] raise exception.PortBindingFailed(port_id=port['id']) [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] nova.exception.PortBindingFailed: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. [ 832.388448] env[63197]: ERROR nova.compute.manager [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] [ 832.388448] env[63197]: DEBUG nova.compute.utils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 832.389613] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.254s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.392400] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Build of instance 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba was re-scheduled: Binding failed for port a742b76d-91fa-4929-8d57-def111bf3bc5, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 832.392815] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 832.393046] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.393194] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.393351] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.634612] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9c27469a-b8f9-4374-ae2a-ca17ed43f4f6] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 832.911205] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.982116] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.095189] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d9907b-bc21-40c3-92c4-9ef766c1be72 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.102846] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca97e08-2525-43e9-9499-dc5a7458c3d9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.133601] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e76d9c-bb86-41da-ab38-1d05c7896ee4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.138050] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.138220] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Cleaning up deleted instances with incomplete migration {{(pid=63197) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 833.142842] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176cfbb4-adab-42d5-ba8a-c7d935e847cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.157013] env[63197]: DEBUG nova.compute.provider_tree [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.484871] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.485141] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 833.485328] env[63197]: DEBUG nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.485494] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.501045] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.643078] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 833.659838] env[63197]: DEBUG nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.004130] env[63197]: DEBUG nova.network.neutron [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.166397] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.775s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.166397] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. [ 834.166397] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Traceback (most recent call last): [ 834.166397] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 834.166397] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self.driver.spawn(context, instance, image_meta, [ 834.166397] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 834.166397] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 834.166397] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 834.166397] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] vm_ref = self.build_virtual_machine(instance, [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] vif_infos = vmwarevif.get_vif_info(self._session, [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] for vif in network_info: [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] return self._sync_wrapper(fn, *args, **kwargs) [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self.wait() [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self[:] = self._gt.wait() [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] return self._exit_event.wait() [ 834.166932] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] current.throw(*self._exc) [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] result = function(*args, **kwargs) [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] return func(*args, **kwargs) [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] raise e [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] nwinfo = self.network_api.allocate_for_instance( [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] created_port_ids = self._update_ports_for_instance( [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 834.167219] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] with excutils.save_and_reraise_exception(): [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] self.force_reraise() [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] raise self.value [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] updated_port = self._update_port( [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] _ensure_no_port_binding_failure(port) [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] raise exception.PortBindingFailed(port_id=port['id']) [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] nova.exception.PortBindingFailed: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. [ 834.167499] env[63197]: ERROR nova.compute.manager [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] [ 834.167745] env[63197]: DEBUG nova.compute.utils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 834.167745] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Build of instance 7786e5f4-a7c9-4c14-b407-d2eda85bf37a was re-scheduled: Binding failed for port 0c4c40f4-2c3a-412d-b444-187d6badf45a, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 834.168170] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 834.168400] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.168547] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquired lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.168703] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.170061] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.684s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.171515] env[63197]: INFO nova.compute.claims [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.507019] env[63197]: INFO nova.compute.manager [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba] Took 1.02 seconds to deallocate network for instance. [ 834.691579] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.776422] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.278652] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Releasing lock "refresh_cache-7786e5f4-a7c9-4c14-b407-d2eda85bf37a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.278885] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 835.279075] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.279249] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.296268] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.458207] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc87703-ce03-4ad2-a128-6a4ce71759a3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.465628] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f60e775-6b65-4e92-a337-a66cd99b10fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.496917] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fb3a21-704c-43bb-8303-eb8bd05382d9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.504627] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424548b1-c7b2-48f3-a1c3-4fc703a99e97 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.520256] env[63197]: DEBUG nova.compute.provider_tree [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.537165] env[63197]: INFO nova.scheduler.client.report [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted allocations for instance 37a283f9-ceb2-4ca8-ac0b-37d87e9964ba [ 835.799171] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.023170] env[63197]: DEBUG nova.scheduler.client.report [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.047270] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b4276b64-fa92-4888-baaa-1e12238bc0a2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "37a283f9-ceb2-4ca8-ac0b-37d87e9964ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.992s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.301760] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 7786e5f4-a7c9-4c14-b407-d2eda85bf37a] Took 1.02 seconds to deallocate network for instance. [ 836.528933] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.529603] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.532204] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.234s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.533599] env[63197]: INFO nova.compute.claims [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.551498] env[63197]: DEBUG nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 837.038100] env[63197]: DEBUG nova.compute.utils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.041674] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.041889] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.069959] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.103012] env[63197]: DEBUG nova.policy [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '703be05550ae4740945ac250d91e3328', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ba36c2f1370406fb8a5c15fbbc42e53', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.339008] env[63197]: INFO nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Deleted allocations for instance 7786e5f4-a7c9-4c14-b407-d2eda85bf37a [ 837.384720] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Successfully created port: da9482eb-6db9-495d-a949-8aaf83f7e5e2 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.482760] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.483056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.542460] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.755293] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226fc571-af5b-405a-a912-d5e32d402634 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.762653] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb09b8bb-8cfa-45b6-93df-34f4f9b332e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.791400] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-873db0ce-539d-4799-858c-ccbe0a427f2d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.798271] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9a168a-3826-46de-a7c0-860ab3ef74ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.810908] env[63197]: DEBUG nova.compute.provider_tree [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.850503] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "7786e5f4-a7c9-4c14-b407-d2eda85bf37a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.845s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.144144] env[63197]: DEBUG nova.compute.manager [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Received event network-changed-da9482eb-6db9-495d-a949-8aaf83f7e5e2 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.144144] env[63197]: DEBUG nova.compute.manager [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Refreshing instance network info cache due to event network-changed-da9482eb-6db9-495d-a949-8aaf83f7e5e2. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.144144] env[63197]: DEBUG oslo_concurrency.lockutils [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] Acquiring lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.144144] env[63197]: DEBUG oslo_concurrency.lockutils [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] Acquired lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.144144] env[63197]: DEBUG nova.network.neutron [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Refreshing network info cache for port da9482eb-6db9-495d-a949-8aaf83f7e5e2 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.291885] env[63197]: ERROR nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. [ 838.291885] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 838.291885] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 838.291885] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 838.291885] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 838.291885] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 838.291885] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 838.291885] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 838.291885] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 838.291885] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 838.291885] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 838.291885] env[63197]: ERROR nova.compute.manager raise self.value [ 838.291885] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 838.291885] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 838.291885] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 838.291885] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 838.292339] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 838.292339] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 838.292339] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. [ 838.292339] env[63197]: ERROR nova.compute.manager [ 838.292339] env[63197]: Traceback (most recent call last): [ 838.292339] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 838.292339] env[63197]: listener.cb(fileno) [ 838.292339] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 838.292339] env[63197]: result = function(*args, **kwargs) [ 838.292339] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 838.292339] env[63197]: return func(*args, **kwargs) [ 838.292339] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 838.292339] env[63197]: raise e [ 838.292339] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 838.292339] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 838.292339] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 838.292339] env[63197]: created_port_ids = self._update_ports_for_instance( [ 838.292339] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 838.292339] env[63197]: with excutils.save_and_reraise_exception(): [ 838.292339] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 838.292339] env[63197]: self.force_reraise() [ 838.292339] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 838.292339] env[63197]: raise self.value [ 838.292339] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 838.292339] env[63197]: updated_port = self._update_port( [ 838.292339] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 838.292339] env[63197]: _ensure_no_port_binding_failure(port) [ 838.292339] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 838.292339] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 838.292996] env[63197]: nova.exception.PortBindingFailed: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. [ 838.292996] env[63197]: Removing descriptor: 18 [ 838.314211] env[63197]: DEBUG nova.scheduler.client.report [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.354576] env[63197]: DEBUG nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.554990] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.579225] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.579583] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.579819] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.580102] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.580333] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.580562] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.580848] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.581096] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.581348] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.581599] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.581859] env[63197]: DEBUG nova.virt.hardware [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.582779] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b17f2b1f-b7cf-467d-8d51-6322b20bae6c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.590827] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6152ac-e4fb-4ea0-87ea-b6038b16c2c9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.604266] env[63197]: ERROR nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Traceback (most recent call last): [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] yield resources [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self.driver.spawn(context, instance, image_meta, [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] vm_ref = self.build_virtual_machine(instance, [ 838.604266] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] for vif in network_info: [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] return self._sync_wrapper(fn, *args, **kwargs) [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self.wait() [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self[:] = self._gt.wait() [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] return self._exit_event.wait() [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 838.604823] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] current.throw(*self._exc) [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] result = function(*args, **kwargs) [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] return func(*args, **kwargs) [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] raise e [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] nwinfo = self.network_api.allocate_for_instance( [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] created_port_ids = self._update_ports_for_instance( [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] with excutils.save_and_reraise_exception(): [ 838.605533] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self.force_reraise() [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] raise self.value [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] updated_port = self._update_port( [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] _ensure_no_port_binding_failure(port) [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] raise exception.PortBindingFailed(port_id=port['id']) [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] nova.exception.PortBindingFailed: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. [ 838.606105] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] [ 838.606105] env[63197]: INFO nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Terminating instance [ 838.609107] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Acquiring lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.663063] env[63197]: DEBUG nova.network.neutron [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.744587] env[63197]: DEBUG nova.network.neutron [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.819367] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.819911] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.822407] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.684s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.824223] env[63197]: INFO nova.compute.claims [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.879285] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.249991] env[63197]: DEBUG oslo_concurrency.lockutils [req-fff290aa-cdbd-4255-91f9-abee95970384 req-f9beaedc-f35e-4e72-96bf-b6cc20d6f56b service nova] Releasing lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.249991] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Acquired lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.249991] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.330289] env[63197]: DEBUG nova.compute.utils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.331994] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.332168] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 839.380605] env[63197]: DEBUG nova.policy [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c91dabb95115428cac178b75dbd4a8ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3e2f93ce59984a2f8ced7987b99df5cb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.655096] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Successfully created port: cd260090-5f70-4fb0-8355-d52e82eb44f9 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.769543] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.835828] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.853259] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.073652] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5898bf8d-5249-4cd5-9410-6e06be25adcd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.081192] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72dc542-e18e-42d1-9059-b1b08343c6c4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.111892] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0e3686-7fc1-45f7-abda-705019667b3b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.118916] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23901bd4-6fd6-49cd-afaa-b6cbf8f3bb1c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.131666] env[63197]: DEBUG nova.compute.provider_tree [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.167605] env[63197]: DEBUG nova.compute.manager [req-f5043ab1-5107-4f67-9667-12a3298c3af1 req-d2a07ac6-9f93-4a98-8000-a1dbeebdff2b service nova] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Received event network-vif-deleted-da9482eb-6db9-495d-a949-8aaf83f7e5e2 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.358087] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Releasing lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.358476] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.358669] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.358959] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-110003d6-bdf5-4b4f-b61e-f11f367d2440 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.368373] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caf5843-dee6-4baf-addf-e4fc34567487 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.391811] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fc4d4223-b0cb-4a40-bd5f-62ec740b57d3 could not be found. [ 840.392090] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.392209] env[63197]: INFO nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Took 0.03 seconds to destroy the instance on the hypervisor. [ 840.392458] env[63197]: DEBUG oslo.service.loopingcall [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.392695] env[63197]: DEBUG nova.compute.manager [-] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.392818] env[63197]: DEBUG nova.network.neutron [-] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.407211] env[63197]: DEBUG nova.network.neutron [-] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.564606] env[63197]: ERROR nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. [ 840.564606] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 840.564606] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 840.564606] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 840.564606] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 840.564606] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 840.564606] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 840.564606] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 840.564606] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 840.564606] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 840.564606] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 840.564606] env[63197]: ERROR nova.compute.manager raise self.value [ 840.564606] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 840.564606] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 840.564606] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 840.564606] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 840.565033] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 840.565033] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 840.565033] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. [ 840.565033] env[63197]: ERROR nova.compute.manager [ 840.565033] env[63197]: Traceback (most recent call last): [ 840.565033] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 840.565033] env[63197]: listener.cb(fileno) [ 840.565033] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 840.565033] env[63197]: result = function(*args, **kwargs) [ 840.565033] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 840.565033] env[63197]: return func(*args, **kwargs) [ 840.565033] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 840.565033] env[63197]: raise e [ 840.565033] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 840.565033] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 840.565033] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 840.565033] env[63197]: created_port_ids = self._update_ports_for_instance( [ 840.565033] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 840.565033] env[63197]: with excutils.save_and_reraise_exception(): [ 840.565033] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 840.565033] env[63197]: self.force_reraise() [ 840.565033] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 840.565033] env[63197]: raise self.value [ 840.565033] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 840.565033] env[63197]: updated_port = self._update_port( [ 840.565033] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 840.565033] env[63197]: _ensure_no_port_binding_failure(port) [ 840.565033] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 840.565033] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 840.565672] env[63197]: nova.exception.PortBindingFailed: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. [ 840.565672] env[63197]: Removing descriptor: 18 [ 840.634790] env[63197]: DEBUG nova.scheduler.client.report [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.844711] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.868670] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.868931] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.869100] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.869285] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.869433] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.869612] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.869824] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.870033] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.870156] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.870316] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.870481] env[63197]: DEBUG nova.virt.hardware [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.871339] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41693be1-dd6e-497f-8fbb-774389d5236c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.878710] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fbac31-db2f-4c5d-a3f0-1e4be58ad50d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.892837] env[63197]: ERROR nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Traceback (most recent call last): [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] yield resources [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self.driver.spawn(context, instance, image_meta, [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] vm_ref = self.build_virtual_machine(instance, [ 840.892837] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] vif_infos = vmwarevif.get_vif_info(self._session, [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] for vif in network_info: [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] return self._sync_wrapper(fn, *args, **kwargs) [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self.wait() [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self[:] = self._gt.wait() [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] return self._exit_event.wait() [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 840.893245] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] current.throw(*self._exc) [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] result = function(*args, **kwargs) [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] return func(*args, **kwargs) [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] raise e [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] nwinfo = self.network_api.allocate_for_instance( [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] created_port_ids = self._update_ports_for_instance( [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] with excutils.save_and_reraise_exception(): [ 840.893583] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self.force_reraise() [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] raise self.value [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] updated_port = self._update_port( [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] _ensure_no_port_binding_failure(port) [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] raise exception.PortBindingFailed(port_id=port['id']) [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] nova.exception.PortBindingFailed: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. [ 840.893920] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] [ 840.893920] env[63197]: INFO nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Terminating instance [ 840.895097] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.895254] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquired lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.895413] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 840.908963] env[63197]: DEBUG nova.network.neutron [-] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.139920] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.317s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.140472] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 841.143061] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.497s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.412087] env[63197]: INFO nova.compute.manager [-] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Took 1.02 seconds to deallocate network for instance. [ 841.414086] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.415961] env[63197]: DEBUG nova.compute.claims [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 841.416146] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.503445] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.646992] env[63197]: DEBUG nova.compute.utils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.652021] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.652021] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.689406] env[63197]: DEBUG nova.policy [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f43a1f72c46efa2f36553d1a768ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d97b713fdcb4cd58c6bea06ec23a441', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.856297] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e77366e-29a8-41ff-b234-e6d09bc0b36f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.863614] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97bae2d-7c68-45f4-8c37-ce269d45bbad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.892906] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa9ddd2-7b95-448f-af76-6e8dbb974e71 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.899807] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1f3173-15ff-47d1-a620-9eca535214e4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.913512] env[63197]: DEBUG nova.compute.provider_tree [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.959343] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Successfully created port: 7455e2c7-9551-4a3b-b203-818e965b59ad {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.006405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Releasing lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.006833] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 842.007038] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 842.007323] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-812288fa-00c3-4e43-91ec-e67918eec9f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.742904] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 842.745681] env[63197]: DEBUG nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.749178] env[63197]: DEBUG nova.compute.manager [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Received event network-changed-cd260090-5f70-4fb0-8355-d52e82eb44f9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.749357] env[63197]: DEBUG nova.compute.manager [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Refreshing instance network info cache due to event network-changed-cd260090-5f70-4fb0-8355-d52e82eb44f9. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.749558] env[63197]: DEBUG oslo_concurrency.lockutils [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] Acquiring lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.749697] env[63197]: DEBUG oslo_concurrency.lockutils [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] Acquired lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.749852] env[63197]: DEBUG nova.network.neutron [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Refreshing network info cache for port cd260090-5f70-4fb0-8355-d52e82eb44f9 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.758225] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb9cabc-e375-46e6-88c5-971339318e79 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.782787] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c005d088-ee27-4d20-9db4-65b47227f22a could not be found. [ 842.783124] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 842.783384] env[63197]: INFO nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Took 0.78 seconds to destroy the instance on the hypervisor. [ 842.783671] env[63197]: DEBUG oslo.service.loopingcall [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.784116] env[63197]: DEBUG nova.compute.manager [-] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 842.784212] env[63197]: DEBUG nova.network.neutron [-] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 842.801652] env[63197]: DEBUG nova.network.neutron [-] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 842.972467] env[63197]: ERROR nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. [ 842.972467] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 842.972467] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 842.972467] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 842.972467] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 842.972467] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 842.972467] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 842.972467] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 842.972467] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 842.972467] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 842.972467] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 842.972467] env[63197]: ERROR nova.compute.manager raise self.value [ 842.972467] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 842.972467] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 842.972467] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 842.972467] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 842.973035] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 842.973035] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 842.973035] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. [ 842.973035] env[63197]: ERROR nova.compute.manager [ 842.973035] env[63197]: Traceback (most recent call last): [ 842.973035] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 842.973035] env[63197]: listener.cb(fileno) [ 842.973035] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 842.973035] env[63197]: result = function(*args, **kwargs) [ 842.973035] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 842.973035] env[63197]: return func(*args, **kwargs) [ 842.973035] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 842.973035] env[63197]: raise e [ 842.973035] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 842.973035] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 842.973035] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 842.973035] env[63197]: created_port_ids = self._update_ports_for_instance( [ 842.973035] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 842.973035] env[63197]: with excutils.save_and_reraise_exception(): [ 842.973035] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 842.973035] env[63197]: self.force_reraise() [ 842.973035] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 842.973035] env[63197]: raise self.value [ 842.973035] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 842.973035] env[63197]: updated_port = self._update_port( [ 842.973035] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 842.973035] env[63197]: _ensure_no_port_binding_failure(port) [ 842.973035] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 842.973035] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 842.973805] env[63197]: nova.exception.PortBindingFailed: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. [ 842.973805] env[63197]: Removing descriptor: 18 [ 843.254975] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.112s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.255280] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Traceback (most recent call last): [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self.driver.spawn(context, instance, image_meta, [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self._vmops.spawn(context, instance, image_meta, injected_files, [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] vm_ref = self.build_virtual_machine(instance, [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] vif_infos = vmwarevif.get_vif_info(self._session, [ 843.255280] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] for vif in network_info: [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] return self._sync_wrapper(fn, *args, **kwargs) [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self.wait() [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self[:] = self._gt.wait() [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] return self._exit_event.wait() [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] current.throw(*self._exc) [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 843.255553] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] result = function(*args, **kwargs) [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] return func(*args, **kwargs) [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] raise e [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] nwinfo = self.network_api.allocate_for_instance( [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] created_port_ids = self._update_ports_for_instance( [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] with excutils.save_and_reraise_exception(): [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] self.force_reraise() [ 843.255833] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] raise self.value [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] updated_port = self._update_port( [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] _ensure_no_port_binding_failure(port) [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] raise exception.PortBindingFailed(port_id=port['id']) [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] nova.exception.PortBindingFailed: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. [ 843.256121] env[63197]: ERROR nova.compute.manager [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] [ 843.256121] env[63197]: DEBUG nova.compute.utils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 843.258924] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.682s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.260316] env[63197]: INFO nova.compute.claims [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.263646] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Build of instance 5128f0c1-da4a-4eb4-9dde-23f830800512 was re-scheduled: Binding failed for port 4bd7fc22-273e-4c8b-a00d-ec7a88087c22, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 843.264241] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 843.264476] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.264621] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquired lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.264773] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.282126] env[63197]: DEBUG nova.network.neutron [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.304237] env[63197]: DEBUG nova.network.neutron [-] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.364204] env[63197]: DEBUG nova.network.neutron [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.759314] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 843.784086] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.784341] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.784496] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.784680] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.784831] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.784958] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.785175] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.785327] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.785487] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.785645] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.785808] env[63197]: DEBUG nova.virt.hardware [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.786660] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b20204d-a6c8-4654-a882-6318a77194b2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.789613] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.796855] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30dac2ae-77fc-4894-9087-1cb31c712a7e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.811217] env[63197]: INFO nova.compute.manager [-] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Took 1.03 seconds to deallocate network for instance. [ 843.812125] env[63197]: ERROR nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Traceback (most recent call last): [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] yield resources [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self.driver.spawn(context, instance, image_meta, [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] vm_ref = self.build_virtual_machine(instance, [ 843.812125] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] for vif in network_info: [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] return self._sync_wrapper(fn, *args, **kwargs) [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self.wait() [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self[:] = self._gt.wait() [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] return self._exit_event.wait() [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 843.812603] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] current.throw(*self._exc) [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] result = function(*args, **kwargs) [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] return func(*args, **kwargs) [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] raise e [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] nwinfo = self.network_api.allocate_for_instance( [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] created_port_ids = self._update_ports_for_instance( [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] with excutils.save_and_reraise_exception(): [ 843.812996] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self.force_reraise() [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] raise self.value [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] updated_port = self._update_port( [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] _ensure_no_port_binding_failure(port) [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] raise exception.PortBindingFailed(port_id=port['id']) [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] nova.exception.PortBindingFailed: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. [ 843.813368] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] [ 843.813368] env[63197]: INFO nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Terminating instance [ 843.815234] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.815387] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.815545] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.817228] env[63197]: DEBUG nova.compute.claims [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 843.817378] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.863345] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.866059] env[63197]: DEBUG oslo_concurrency.lockutils [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] Releasing lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.866281] env[63197]: DEBUG nova.compute.manager [req-093e3ff8-0fe7-453f-9fad-d9c41033d528 req-98e37ab1-79fe-4489-b266-729819a0354f service nova] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Received event network-vif-deleted-cd260090-5f70-4fb0-8355-d52e82eb44f9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.222953] env[63197]: DEBUG nova.compute.manager [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Received event network-changed-7455e2c7-9551-4a3b-b203-818e965b59ad {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.223166] env[63197]: DEBUG nova.compute.manager [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Refreshing instance network info cache due to event network-changed-7455e2c7-9551-4a3b-b203-818e965b59ad. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.223361] env[63197]: DEBUG oslo_concurrency.lockutils [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] Acquiring lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.333369] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.365630] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Releasing lock "refresh_cache-5128f0c1-da4a-4eb4-9dde-23f830800512" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.365863] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 844.366049] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.366217] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.381198] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.416488] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.475121] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129aa66e-82f8-45b9-bb93-fd792f60cec1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.482935] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c16ff9-8666-4ff7-af82-2c9fc1b56184 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.511813] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6762db-c81e-4fd8-bf77-3cb63dcbd576 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.518437] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa890ac-5a1c-48a9-bf4f-c0646ab7f5a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.530955] env[63197]: DEBUG nova.compute.provider_tree [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.885629] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.919280] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.919652] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.919833] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.920126] env[63197]: DEBUG oslo_concurrency.lockutils [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] Acquired lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.920287] env[63197]: DEBUG nova.network.neutron [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Refreshing network info cache for port 7455e2c7-9551-4a3b-b203-818e965b59ad {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.921150] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a783da34-ee52-41b6-9173-a8deb0e0b609 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.930835] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07686e1-7061-475f-8464-9b59b2f33188 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.951170] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 643622a1-cb00-4829-a83a-a347e6e50e5d could not be found. [ 844.951361] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 844.951554] env[63197]: INFO nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 844.951791] env[63197]: DEBUG oslo.service.loopingcall [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.951991] env[63197]: DEBUG nova.compute.manager [-] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.952096] env[63197]: DEBUG nova.network.neutron [-] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.967266] env[63197]: DEBUG nova.network.neutron [-] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.034805] env[63197]: DEBUG nova.scheduler.client.report [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.390047] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 5128f0c1-da4a-4eb4-9dde-23f830800512] Took 1.02 seconds to deallocate network for instance. [ 845.440991] env[63197]: DEBUG nova.network.neutron [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.469776] env[63197]: DEBUG nova.network.neutron [-] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.517985] env[63197]: DEBUG nova.network.neutron [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.539793] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.281s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.540309] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.542886] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.134s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.971805] env[63197]: INFO nova.compute.manager [-] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Took 1.02 seconds to deallocate network for instance. [ 845.974104] env[63197]: DEBUG nova.compute.claims [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 845.974281] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.020223] env[63197]: DEBUG oslo_concurrency.lockutils [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] Releasing lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.020473] env[63197]: DEBUG nova.compute.manager [req-a0e2dd46-dff6-4c12-8c13-c82b8a43c31b req-afb0fb7e-50d4-4957-85fa-09cf0a5e8b09 service nova] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Received event network-vif-deleted-7455e2c7-9551-4a3b-b203-818e965b59ad {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.047010] env[63197]: DEBUG nova.compute.utils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.051086] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.051256] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 846.096433] env[63197]: DEBUG nova.policy [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12d00044b65c48f484ac9fea595d6179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80138785cfdb4a2188ffb972da437646', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.244610] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fac1db-a126-446b-8ff0-d18693f81db7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.252469] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35d3ed7-fc22-4c27-a48f-70c34be53bd1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.282258] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e1c5bf-4220-4bfb-a82c-ad2391d19a1c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.289400] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1165a492-d568-4c71-be72-a703632c61de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.303485] env[63197]: DEBUG nova.compute.provider_tree [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.364982] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Successfully created port: a31e966e-24dd-4793-8bb1-fe856caf9242 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.418260] env[63197]: INFO nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Deleted allocations for instance 5128f0c1-da4a-4eb4-9dde-23f830800512 [ 846.553973] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.807059] env[63197]: DEBUG nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.930760] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "5128f0c1-da4a-4eb4-9dde-23f830800512" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.895s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.143370] env[63197]: DEBUG nova.compute.manager [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Received event network-changed-a31e966e-24dd-4793-8bb1-fe856caf9242 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.143757] env[63197]: DEBUG nova.compute.manager [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Refreshing instance network info cache due to event network-changed-a31e966e-24dd-4793-8bb1-fe856caf9242. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.143987] env[63197]: DEBUG oslo_concurrency.lockutils [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] Acquiring lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.144173] env[63197]: DEBUG oslo_concurrency.lockutils [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] Acquired lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.144360] env[63197]: DEBUG nova.network.neutron [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Refreshing network info cache for port a31e966e-24dd-4793-8bb1-fe856caf9242 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.285419] env[63197]: ERROR nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. [ 847.285419] env[63197]: ERROR nova.compute.manager Traceback (most recent call last): [ 847.285419] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 847.285419] env[63197]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 847.285419] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 847.285419] env[63197]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 847.285419] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 847.285419] env[63197]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 847.285419] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 847.285419] env[63197]: ERROR nova.compute.manager self.force_reraise() [ 847.285419] env[63197]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 847.285419] env[63197]: ERROR nova.compute.manager raise self.value [ 847.285419] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 847.285419] env[63197]: ERROR nova.compute.manager updated_port = self._update_port( [ 847.285419] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 847.285419] env[63197]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 847.286185] env[63197]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 847.286185] env[63197]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 847.286185] env[63197]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. [ 847.286185] env[63197]: ERROR nova.compute.manager [ 847.286185] env[63197]: Traceback (most recent call last): [ 847.286185] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 847.286185] env[63197]: listener.cb(fileno) [ 847.286185] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 847.286185] env[63197]: result = function(*args, **kwargs) [ 847.286185] env[63197]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 847.286185] env[63197]: return func(*args, **kwargs) [ 847.286185] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 847.286185] env[63197]: raise e [ 847.286185] env[63197]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 847.286185] env[63197]: nwinfo = self.network_api.allocate_for_instance( [ 847.286185] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 847.286185] env[63197]: created_port_ids = self._update_ports_for_instance( [ 847.286185] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 847.286185] env[63197]: with excutils.save_and_reraise_exception(): [ 847.286185] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 847.286185] env[63197]: self.force_reraise() [ 847.286185] env[63197]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 847.286185] env[63197]: raise self.value [ 847.286185] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 847.286185] env[63197]: updated_port = self._update_port( [ 847.286185] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 847.286185] env[63197]: _ensure_no_port_binding_failure(port) [ 847.286185] env[63197]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 847.286185] env[63197]: raise exception.PortBindingFailed(port_id=port['id']) [ 847.287737] env[63197]: nova.exception.PortBindingFailed: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. [ 847.287737] env[63197]: Removing descriptor: 18 [ 847.312828] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.770s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.313940] env[63197]: ERROR nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Traceback (most recent call last): [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self.driver.spawn(context, instance, image_meta, [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self._vmops.spawn(context, instance, image_meta, injected_files, [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] vm_ref = self.build_virtual_machine(instance, [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] vif_infos = vmwarevif.get_vif_info(self._session, [ 847.313940] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] for vif in network_info: [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return self._sync_wrapper(fn, *args, **kwargs) [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self.wait() [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self[:] = self._gt.wait() [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return self._exit_event.wait() [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] result = hub.switch() [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 847.314453] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return self.greenlet.switch() [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] result = function(*args, **kwargs) [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] return func(*args, **kwargs) [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] raise e [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] nwinfo = self.network_api.allocate_for_instance( [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] created_port_ids = self._update_ports_for_instance( [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] with excutils.save_and_reraise_exception(): [ 847.314852] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] self.force_reraise() [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] raise self.value [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] updated_port = self._update_port( [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] _ensure_no_port_binding_failure(port) [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] raise exception.PortBindingFailed(port_id=port['id']) [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] nova.exception.PortBindingFailed: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. [ 847.315225] env[63197]: ERROR nova.compute.manager [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] [ 847.316090] env[63197]: DEBUG nova.compute.utils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 847.316090] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.775s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.316966] env[63197]: INFO nova.compute.claims [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.319542] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Build of instance 06167654-0ac8-4b04-8eb7-071eac0894ae was re-scheduled: Binding failed for port 7475c740-8ca9-47dd-a390-89fd8189b300, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 847.319971] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 847.320227] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquiring lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.320373] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Acquired lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.320529] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 847.430738] env[63197]: DEBUG nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.562649] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.587347] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.587610] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.587763] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.587945] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.588107] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.588255] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.588455] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.588610] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.588774] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.589030] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.589125] env[63197]: DEBUG nova.virt.hardware [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.590079] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c369b1f9-8ceb-4629-9127-99badce54744 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.598511] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7c357d-d217-418f-ab21-ebf4607496e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.612494] env[63197]: ERROR nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Traceback (most recent call last): [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] yield resources [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self.driver.spawn(context, instance, image_meta, [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] vm_ref = self.build_virtual_machine(instance, [ 847.612494] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] for vif in network_info: [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] return self._sync_wrapper(fn, *args, **kwargs) [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self.wait() [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self[:] = self._gt.wait() [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] return self._exit_event.wait() [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 847.612900] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] current.throw(*self._exc) [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] result = function(*args, **kwargs) [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] return func(*args, **kwargs) [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] raise e [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] nwinfo = self.network_api.allocate_for_instance( [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] created_port_ids = self._update_ports_for_instance( [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] with excutils.save_and_reraise_exception(): [ 847.613281] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self.force_reraise() [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] raise self.value [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] updated_port = self._update_port( [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] _ensure_no_port_binding_failure(port) [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] raise exception.PortBindingFailed(port_id=port['id']) [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] nova.exception.PortBindingFailed: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. [ 847.613653] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] [ 847.613653] env[63197]: INFO nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Terminating instance [ 847.614782] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.662455] env[63197]: DEBUG nova.network.neutron [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.746467] env[63197]: DEBUG nova.network.neutron [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.843437] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 847.934174] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.952263] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.249242] env[63197]: DEBUG oslo_concurrency.lockutils [req-58899e66-d5ea-48d3-98c3-a6ec15e53dc2 req-e9736870-c286-4cef-827b-f1921e2236fd service nova] Releasing lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.249692] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.249873] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.439505] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Releasing lock "refresh_cache-06167654-0ac8-4b04-8eb7-071eac0894ae" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.439748] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 848.439927] env[63197]: DEBUG nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 848.440112] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 848.455302] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.521655] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b6c6de-2599-4076-8678-3c9bbbdd43f3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.529262] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457bca96-9ed2-470a-b5c0-eddfa6b23348 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.558536] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002ebf5b-cd51-4e3a-af6b-2aa6c779889e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.565105] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8e5954-6bf9-457d-9c05-9ed10352d889 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.577412] env[63197]: DEBUG nova.compute.provider_tree [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.766383] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.844785] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.957548] env[63197]: DEBUG nova.network.neutron [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.080065] env[63197]: DEBUG nova.scheduler.client.report [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.169500] env[63197]: DEBUG nova.compute.manager [req-5472414f-e39a-4457-a9a4-010d6e1f1fa0 req-007b2487-903d-4c53-bbc2-0fa0b6cbe242 service nova] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Received event network-vif-deleted-a31e966e-24dd-4793-8bb1-fe856caf9242 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.347246] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.347630] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 849.347817] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 849.348120] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c01799c-88fe-4f65-adb1-5483fe9a305b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.357494] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb49b119-4555-4776-9437-2f3c9fe28c46 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.377449] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92ee8e7a-c169-44d8-8f74-84590b4f62e9 could not be found. [ 849.377648] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 849.377823] env[63197]: INFO nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 849.378067] env[63197]: DEBUG oslo.service.loopingcall [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.378279] env[63197]: DEBUG nova.compute.manager [-] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.378679] env[63197]: DEBUG nova.network.neutron [-] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 849.392724] env[63197]: DEBUG nova.network.neutron [-] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.460481] env[63197]: INFO nova.compute.manager [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] [instance: 06167654-0ac8-4b04-8eb7-071eac0894ae] Took 1.02 seconds to deallocate network for instance. [ 849.584914] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.585507] env[63197]: DEBUG nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.591183] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.391s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.895767] env[63197]: DEBUG nova.network.neutron [-] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.096485] env[63197]: DEBUG nova.compute.utils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.100546] env[63197]: DEBUG nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 850.100713] env[63197]: DEBUG nova.network.neutron [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.159829] env[63197]: DEBUG nova.policy [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ccf6464e6854969a32f2388f7866834', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '558e5440e64c43f59d0f2cc54c18c79f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.374133] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769a021e-0e6c-40a4-9dcf-1c919002ac25 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.381983] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a16336-b81d-4a5a-b1cc-a1d49eeeb926 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.418487] env[63197]: INFO nova.compute.manager [-] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Took 1.04 seconds to deallocate network for instance. [ 850.418487] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf1cc53-76e8-450c-be13-369d8e8d6dbd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.421120] env[63197]: DEBUG nova.compute.claims [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Aborting claim: {{(pid=63197) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 850.421311] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.427111] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72480c57-cbfd-484b-a092-8e56681cda5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.444235] env[63197]: DEBUG nova.compute.provider_tree [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.446173] env[63197]: DEBUG nova.network.neutron [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Successfully created port: 5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 850.499166] env[63197]: INFO nova.scheduler.client.report [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Deleted allocations for instance 06167654-0ac8-4b04-8eb7-071eac0894ae [ 850.604152] env[63197]: DEBUG nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.949830] env[63197]: DEBUG nova.scheduler.client.report [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.010829] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ddba7fef-acca-49ba-bbe3-d83bfa327b73 tempest-ListServersNegativeTestJSON-1415834991 tempest-ListServersNegativeTestJSON-1415834991-project-member] Lock "06167654-0ac8-4b04-8eb7-071eac0894ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.948s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.454966] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.864s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.455614] env[63197]: ERROR nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Traceback (most recent call last): [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self.driver.spawn(context, instance, image_meta, [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self._vmops.spawn(context, instance, image_meta, injected_files, [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] vm_ref = self.build_virtual_machine(instance, [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] vif_infos = vmwarevif.get_vif_info(self._session, [ 851.455614] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] for vif in network_info: [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] return self._sync_wrapper(fn, *args, **kwargs) [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self.wait() [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self[:] = self._gt.wait() [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] return self._exit_event.wait() [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] current.throw(*self._exc) [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 851.456333] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] result = function(*args, **kwargs) [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] return func(*args, **kwargs) [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] raise e [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] nwinfo = self.network_api.allocate_for_instance( [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] created_port_ids = self._update_ports_for_instance( [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] with excutils.save_and_reraise_exception(): [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] self.force_reraise() [ 851.456845] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] raise self.value [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] updated_port = self._update_port( [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] _ensure_no_port_binding_failure(port) [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] raise exception.PortBindingFailed(port_id=port['id']) [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] nova.exception.PortBindingFailed: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. [ 851.457357] env[63197]: ERROR nova.compute.manager [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] [ 851.457357] env[63197]: DEBUG nova.compute.utils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 851.457749] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.388s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.459011] env[63197]: INFO nova.compute.claims [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.461833] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Build of instance f8a0c645-4595-41e7-b564-2f74fed9f275 was re-scheduled: Binding failed for port 4570b946-20eb-46f8-ad69-d9cfc02035fa, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 851.462256] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 851.462477] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquiring lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.462625] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Acquired lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.462779] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.514794] env[63197]: DEBUG nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.614775] env[63197]: DEBUG nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.639291] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.639500] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.639788] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.639986] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.640159] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.640308] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.640517] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.640672] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.640948] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.641034] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.641240] env[63197]: DEBUG nova.virt.hardware [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.643158] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b7bd74-e422-47d2-afd4-8b551e932ce9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.651030] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd511c4-e1b1-47d7-99d5-50302cc59207 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.988384] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.033426] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.077540] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.200737] env[63197]: DEBUG nova.compute.manager [req-ecd22896-dd26-4609-8d74-bd1f10bba447 req-18156ab5-ebf5-4041-bf7b-e1efdb656b2e service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received event network-vif-plugged-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.200939] env[63197]: DEBUG oslo_concurrency.lockutils [req-ecd22896-dd26-4609-8d74-bd1f10bba447 req-18156ab5-ebf5-4041-bf7b-e1efdb656b2e service nova] Acquiring lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.201169] env[63197]: DEBUG oslo_concurrency.lockutils [req-ecd22896-dd26-4609-8d74-bd1f10bba447 req-18156ab5-ebf5-4041-bf7b-e1efdb656b2e service nova] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.201337] env[63197]: DEBUG oslo_concurrency.lockutils [req-ecd22896-dd26-4609-8d74-bd1f10bba447 req-18156ab5-ebf5-4041-bf7b-e1efdb656b2e service nova] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.201496] env[63197]: DEBUG nova.compute.manager [req-ecd22896-dd26-4609-8d74-bd1f10bba447 req-18156ab5-ebf5-4041-bf7b-e1efdb656b2e service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] No waiting events found dispatching network-vif-plugged-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.201716] env[63197]: WARNING nova.compute.manager [req-ecd22896-dd26-4609-8d74-bd1f10bba447 req-18156ab5-ebf5-4041-bf7b-e1efdb656b2e service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received unexpected event network-vif-plugged-5f405bfb-7d3e-4876-8fe0-2dee955c3325 for instance with vm_state building and task_state spawning. [ 852.410437] env[63197]: DEBUG nova.network.neutron [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Successfully updated port: 5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.580950] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Releasing lock "refresh_cache-f8a0c645-4595-41e7-b564-2f74fed9f275" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.581230] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 852.581414] env[63197]: DEBUG nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 852.581623] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 852.602014] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.687046] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12956824-29f4-4637-a660-af67323f7424 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.694243] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc75b24-24f6-4abc-b73f-30d064cb4163 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.731536] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efda4e3c-766f-4763-8236-f1c9edf7ebe2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.740252] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9d2db8-1cde-4ec4-9df2-e3786b977730 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.756970] env[63197]: DEBUG nova.compute.provider_tree [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.913799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.913799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.913799] env[63197]: DEBUG nova.network.neutron [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.105021] env[63197]: DEBUG nova.network.neutron [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.260691] env[63197]: DEBUG nova.scheduler.client.report [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.444874] env[63197]: DEBUG nova.network.neutron [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.610731] env[63197]: INFO nova.compute.manager [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] [instance: f8a0c645-4595-41e7-b564-2f74fed9f275] Took 1.03 seconds to deallocate network for instance. [ 853.659217] env[63197]: DEBUG nova.network.neutron [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.766021] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.766565] env[63197]: DEBUG nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.769157] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.890s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.770666] env[63197]: INFO nova.compute.claims [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.162128] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Releasing lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.162459] env[63197]: DEBUG nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Instance network_info: |[{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.162953] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:df:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee4b2432-c393-4e50-ae0e-b5e12bad37db', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f405bfb-7d3e-4876-8fe0-2dee955c3325', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.170577] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Creating folder: Project (558e5440e64c43f59d0f2cc54c18c79f). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.171073] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b8ac9411-820d-469c-94b2-9e4cff2bf246 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.181849] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Created folder: Project (558e5440e64c43f59d0f2cc54c18c79f) in parent group-v290286. [ 854.182045] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Creating folder: Instances. Parent ref: group-v290306. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.182274] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-777ce74d-a254-4db8-a1e3-2c391a5556af {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.190986] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Created folder: Instances in parent group-v290306. [ 854.191229] env[63197]: DEBUG oslo.service.loopingcall [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.191413] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.191632] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed74ebd7-97bd-46bc-9353-5f11fc0c3b08 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.208725] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.208725] env[63197]: value = "task-1364097" [ 854.208725] env[63197]: _type = "Task" [ 854.208725] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.215947] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364097, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.225265] env[63197]: DEBUG nova.compute.manager [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.225462] env[63197]: DEBUG nova.compute.manager [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing instance network info cache due to event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.225733] env[63197]: DEBUG oslo_concurrency.lockutils [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] Acquiring lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.225872] env[63197]: DEBUG oslo_concurrency.lockutils [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] Acquired lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.226043] env[63197]: DEBUG nova.network.neutron [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.275582] env[63197]: DEBUG nova.compute.utils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.279650] env[63197]: DEBUG nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.279928] env[63197]: DEBUG nova.network.neutron [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 854.335965] env[63197]: DEBUG nova.policy [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e8b7c8ad97b4a3eb86ba7c7275a5b6e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '185f0eaa93b5435295800f0c7642cb4e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.658373] env[63197]: INFO nova.scheduler.client.report [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Deleted allocations for instance f8a0c645-4595-41e7-b564-2f74fed9f275 [ 854.668088] env[63197]: DEBUG nova.network.neutron [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Successfully created port: 375c2467-68af-4a0b-8cce-2209b198b1d8 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.726020] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364097, 'name': CreateVM_Task, 'duration_secs': 0.30832} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.730452] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.744187] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.744520] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.744928] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.746013] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a172c51-03ee-4bb7-a544-087b1d96b25e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.754511] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 854.754511] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52cb9229-7136-8bcb-52bf-51c615d6f93b" [ 854.754511] env[63197]: _type = "Task" [ 854.754511] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.764102] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52cb9229-7136-8bcb-52bf-51c615d6f93b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.781635] env[63197]: DEBUG nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.136114] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44dacf8a-11d2-46b8-a870-16ff97f004be {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.145515] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8c9f60-9363-427e-b44d-d2b68416ce56 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.184230] env[63197]: DEBUG oslo_concurrency.lockutils [None req-490f8377-2521-4338-a2e3-746391cde5a3 tempest-ServerGroupTestJSON-830473531 tempest-ServerGroupTestJSON-830473531-project-member] Lock "f8a0c645-4595-41e7-b564-2f74fed9f275" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.496s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.187055] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3d54c5-cf5c-4163-9a46-6f48ce0176f0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.198934] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26df8df-1be4-47c9-92de-257c8227c2ec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.213694] env[63197]: DEBUG nova.compute.provider_tree [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.264520] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52cb9229-7136-8bcb-52bf-51c615d6f93b, 'name': SearchDatastore_Task, 'duration_secs': 0.010642} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.264593] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.264822] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.265063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.265208] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.265382] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.265628] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abb02524-85be-4140-8a8c-fce006a4025f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.273249] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.273425] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.274188] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e62c4f7-d41a-40c0-b79b-8a912e42e3e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.279099] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 855.279099] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ca79f4-9dcb-ee60-6bf2-ffb2f01fc9f0" [ 855.279099] env[63197]: _type = "Task" [ 855.279099] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.286339] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ca79f4-9dcb-ee60-6bf2-ffb2f01fc9f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.295925] env[63197]: DEBUG nova.network.neutron [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updated VIF entry in instance network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.296268] env[63197]: DEBUG nova.network.neutron [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.692999] env[63197]: DEBUG nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.719024] env[63197]: DEBUG nova.scheduler.client.report [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.795127] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ca79f4-9dcb-ee60-6bf2-ffb2f01fc9f0, 'name': SearchDatastore_Task, 'duration_secs': 0.008437} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.796202] env[63197]: DEBUG nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.800724] env[63197]: DEBUG oslo_concurrency.lockutils [req-43c95a32-8d38-4f9d-bc59-6304750c0491 req-cde17e6a-a891-4352-828f-77a39b8dc09b service nova] Releasing lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.800724] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40d07392-02aa-4992-8489-9c7cde477a4f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.806026] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 855.806026] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a88049-67c2-0324-5447-32034b7a0902" [ 855.806026] env[63197]: _type = "Task" [ 855.806026] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.815095] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a88049-67c2-0324-5447-32034b7a0902, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.829139] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.829393] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.829548] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.829728] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.829871] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.830023] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.830233] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.830391] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.831029] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.831029] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.831029] env[63197]: DEBUG nova.virt.hardware [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.831740] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9af2e9-7496-48d1-badb-d1439268de53 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.839385] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5fe53d-6814-40f9-ba5f-a424c106b2f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.218102] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.227958] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.228120] env[63197]: DEBUG nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 856.231737] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.815s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.264624] env[63197]: DEBUG nova.network.neutron [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Successfully updated port: 375c2467-68af-4a0b-8cce-2209b198b1d8 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 856.290135] env[63197]: DEBUG nova.compute.manager [req-f6b6ba5b-04b6-4154-9fb3-b7a2a565cbfd req-b69be62a-5ce7-450f-bab0-b5daa5ebcfd5 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Received event network-vif-plugged-375c2467-68af-4a0b-8cce-2209b198b1d8 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.290356] env[63197]: DEBUG oslo_concurrency.lockutils [req-f6b6ba5b-04b6-4154-9fb3-b7a2a565cbfd req-b69be62a-5ce7-450f-bab0-b5daa5ebcfd5 service nova] Acquiring lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.290566] env[63197]: DEBUG oslo_concurrency.lockutils [req-f6b6ba5b-04b6-4154-9fb3-b7a2a565cbfd req-b69be62a-5ce7-450f-bab0-b5daa5ebcfd5 service nova] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.290735] env[63197]: DEBUG oslo_concurrency.lockutils [req-f6b6ba5b-04b6-4154-9fb3-b7a2a565cbfd req-b69be62a-5ce7-450f-bab0-b5daa5ebcfd5 service nova] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.291494] env[63197]: DEBUG nova.compute.manager [req-f6b6ba5b-04b6-4154-9fb3-b7a2a565cbfd req-b69be62a-5ce7-450f-bab0-b5daa5ebcfd5 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] No waiting events found dispatching network-vif-plugged-375c2467-68af-4a0b-8cce-2209b198b1d8 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.291755] env[63197]: WARNING nova.compute.manager [req-f6b6ba5b-04b6-4154-9fb3-b7a2a565cbfd req-b69be62a-5ce7-450f-bab0-b5daa5ebcfd5 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Received unexpected event network-vif-plugged-375c2467-68af-4a0b-8cce-2209b198b1d8 for instance with vm_state building and task_state spawning. [ 856.318244] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a88049-67c2-0324-5447-32034b7a0902, 'name': SearchDatastore_Task, 'duration_secs': 0.009259} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.318499] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.318808] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 65df32fa-a6e7-4c3a-af8b-0422df8d9229/65df32fa-a6e7-4c3a-af8b-0422df8d9229.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.319129] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-79979be5-dd91-4a51-87dc-515662f04e5e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.325986] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 856.325986] env[63197]: value = "task-1364098" [ 856.325986] env[63197]: _type = "Task" [ 856.325986] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.334970] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.740431] env[63197]: DEBUG nova.compute.utils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.749665] env[63197]: DEBUG nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Not allocating networking since 'none' was specified. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 856.771950] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "refresh_cache-c55b5b2b-5ecc-43bb-a279-7370cd9ac722" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.771950] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquired lock "refresh_cache-c55b5b2b-5ecc-43bb-a279-7370cd9ac722" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.772028] env[63197]: DEBUG nova.network.neutron [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.842796] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500215} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.842987] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 65df32fa-a6e7-4c3a-af8b-0422df8d9229/65df32fa-a6e7-4c3a-af8b-0422df8d9229.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.843301] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.843494] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-286102f0-5638-4963-8649-d7020253f699 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.852298] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 856.852298] env[63197]: value = "task-1364099" [ 856.852298] env[63197]: _type = "Task" [ 856.852298] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.864290] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364099, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.009830] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcd1857-e998-4ec8-9354-929e8f184586 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.017430] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec1f52a-85a3-44f8-8402-43c010eb6f5e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.056016] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb04fe4-16e8-4590-8526-c0e68b824127 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.063627] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343c9d21-16b1-4e11-bac6-dc124dd15f83 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.077841] env[63197]: DEBUG nova.compute.provider_tree [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.250359] env[63197]: DEBUG nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 857.324897] env[63197]: DEBUG nova.network.neutron [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 857.363883] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364099, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073582} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.364305] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.364985] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e2f205-8e0d-440e-9b44-923d487ac7b6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.389103] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 65df32fa-a6e7-4c3a-af8b-0422df8d9229/65df32fa-a6e7-4c3a-af8b-0422df8d9229.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.389410] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35b5d945-e33f-499d-963e-b49da86cf9a6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.412019] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 857.412019] env[63197]: value = "task-1364100" [ 857.412019] env[63197]: _type = "Task" [ 857.412019] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.419380] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.555980] env[63197]: DEBUG nova.network.neutron [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Updating instance_info_cache with network_info: [{"id": "375c2467-68af-4a0b-8cce-2209b198b1d8", "address": "fa:16:3e:1f:85:55", "network": {"id": "ae948c4d-6288-4540-b9e5-62f26a6719ae", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1174851250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "185f0eaa93b5435295800f0c7642cb4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap375c2467-68", "ovs_interfaceid": "375c2467-68af-4a0b-8cce-2209b198b1d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.581279] env[63197]: DEBUG nova.scheduler.client.report [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.922518] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364100, 'name': ReconfigVM_Task, 'duration_secs': 0.368764} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.922896] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 65df32fa-a6e7-4c3a-af8b-0422df8d9229/65df32fa-a6e7-4c3a-af8b-0422df8d9229.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.924034] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5426533-03a0-47fe-ba56-94ed0cea6a5a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.936394] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 857.936394] env[63197]: value = "task-1364101" [ 857.936394] env[63197]: _type = "Task" [ 857.936394] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.949398] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364101, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.059378] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Releasing lock "refresh_cache-c55b5b2b-5ecc-43bb-a279-7370cd9ac722" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.059817] env[63197]: DEBUG nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Instance network_info: |[{"id": "375c2467-68af-4a0b-8cce-2209b198b1d8", "address": "fa:16:3e:1f:85:55", "network": {"id": "ae948c4d-6288-4540-b9e5-62f26a6719ae", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1174851250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "185f0eaa93b5435295800f0c7642cb4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap375c2467-68", "ovs_interfaceid": "375c2467-68af-4a0b-8cce-2209b198b1d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.060485] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:85:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2c424c9-6446-4b2a-af8c-4d9c29117c39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '375c2467-68af-4a0b-8cce-2209b198b1d8', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.068843] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Creating folder: Project (185f0eaa93b5435295800f0c7642cb4e). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.069180] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6f327843-11d8-4c1c-aed1-3be99ad97612 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.079352] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Created folder: Project (185f0eaa93b5435295800f0c7642cb4e) in parent group-v290286. [ 858.079601] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Creating folder: Instances. Parent ref: group-v290309. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.079851] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70a097a0-4f2d-45c8-bde0-c14b1c3420f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.086705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.855s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.087363] env[63197]: ERROR nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Traceback (most recent call last): [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self.driver.spawn(context, instance, image_meta, [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] vm_ref = self.build_virtual_machine(instance, [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] vif_infos = vmwarevif.get_vif_info(self._session, [ 858.087363] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] for vif in network_info: [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] return self._sync_wrapper(fn, *args, **kwargs) [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self.wait() [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self[:] = self._gt.wait() [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] return self._exit_event.wait() [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] current.throw(*self._exc) [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 858.087707] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] result = function(*args, **kwargs) [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] return func(*args, **kwargs) [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] raise e [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] nwinfo = self.network_api.allocate_for_instance( [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] created_port_ids = self._update_ports_for_instance( [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] with excutils.save_and_reraise_exception(): [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] self.force_reraise() [ 858.088070] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] raise self.value [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] updated_port = self._update_port( [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] _ensure_no_port_binding_failure(port) [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] raise exception.PortBindingFailed(port_id=port['id']) [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] nova.exception.PortBindingFailed: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. [ 858.088424] env[63197]: ERROR nova.compute.manager [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] [ 858.088424] env[63197]: DEBUG nova.compute.utils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 858.090686] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.273s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.094620] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Created folder: Instances in parent group-v290309. [ 858.094620] env[63197]: DEBUG oslo.service.loopingcall [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.094620] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Build of instance fc4d4223-b0cb-4a40-bd5f-62ec740b57d3 was re-scheduled: Binding failed for port da9482eb-6db9-495d-a949-8aaf83f7e5e2, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 858.094987] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 858.095327] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Acquiring lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.095433] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Acquired lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.096062] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.098334] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.098803] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57e3e504-7e30-4f18-a540-e6fd0e42f0dc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.120694] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.120694] env[63197]: value = "task-1364104" [ 858.120694] env[63197]: _type = "Task" [ 858.120694] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.128925] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364104, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.132433] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.226866] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.263827] env[63197]: DEBUG nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 858.294548] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.294732] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.294979] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.295300] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.295549] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.295800] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.296149] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.296426] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.296754] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.297066] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.297363] env[63197]: DEBUG nova.virt.hardware [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.298405] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22d585a-5b69-46e6-a2b1-59ffee96fc4e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.307167] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0195953d-4911-48b9-b46e-0471278aff11 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.325547] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.331610] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Creating folder: Project (df31d1fd0b454e5fa721015eef2aa0a5). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.333048] env[63197]: DEBUG nova.compute.manager [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Received event network-changed-375c2467-68af-4a0b-8cce-2209b198b1d8 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.333261] env[63197]: DEBUG nova.compute.manager [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Refreshing instance network info cache due to event network-changed-375c2467-68af-4a0b-8cce-2209b198b1d8. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.333507] env[63197]: DEBUG oslo_concurrency.lockutils [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] Acquiring lock "refresh_cache-c55b5b2b-5ecc-43bb-a279-7370cd9ac722" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.333680] env[63197]: DEBUG oslo_concurrency.lockutils [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] Acquired lock "refresh_cache-c55b5b2b-5ecc-43bb-a279-7370cd9ac722" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.333854] env[63197]: DEBUG nova.network.neutron [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Refreshing network info cache for port 375c2467-68af-4a0b-8cce-2209b198b1d8 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.335527] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c383f98-11f1-4a2c-8fbf-2745181b0999 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.345585] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Created folder: Project (df31d1fd0b454e5fa721015eef2aa0a5) in parent group-v290286. [ 858.345888] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Creating folder: Instances. Parent ref: group-v290312. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 858.347181] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cde17c53-7e28-460d-8187-5075add7d7e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.351968] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "638ef9c9-253b-4958-a660-6c1801408a51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.352488] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.359424] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Created folder: Instances in parent group-v290312. [ 858.359677] env[63197]: DEBUG oslo.service.loopingcall [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.359884] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.360219] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f077b11-5c09-4a83-94dc-63e797ebf853 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.385658] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.385658] env[63197]: value = "task-1364107" [ 858.385658] env[63197]: _type = "Task" [ 858.385658] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.393269] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364107, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.447947] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364101, 'name': Rename_Task, 'duration_secs': 0.263943} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.448310] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.448597] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99133c03-1fc8-446f-b0b2-c09aab4d63b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.455124] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 858.455124] env[63197]: value = "task-1364108" [ 858.455124] env[63197]: _type = "Task" [ 858.455124] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.463312] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.633653] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364104, 'name': CreateVM_Task, 'duration_secs': 0.440708} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.634071] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.634986] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.635311] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.635754] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.637820] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc386876-c883-45fc-93a9-a5c05eefa962 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.642333] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 858.642333] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523e7456-0175-81fa-3756-97e6f908ab6e" [ 858.642333] env[63197]: _type = "Task" [ 858.642333] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.654020] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523e7456-0175-81fa-3756-97e6f908ab6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.732020] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Releasing lock "refresh_cache-fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.732020] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 858.732020] env[63197]: DEBUG nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 858.732020] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 858.756665] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.860574] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c01f5ae-172b-4e9e-aeb0-8eb12d7c31d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.868936] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2205eb90-cbd3-4571-9401-56e52db675ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.904141] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142ea71e-2be1-4c7a-a41f-4d1e318c28fe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.915021] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364107, 'name': CreateVM_Task, 'duration_secs': 0.280297} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.917262] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.917783] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.918987] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1fcc375-e9dd-4c12-8671-6149941308b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.933012] env[63197]: DEBUG nova.compute.provider_tree [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.970783] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364108, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.154544] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523e7456-0175-81fa-3756-97e6f908ab6e, 'name': SearchDatastore_Task, 'duration_secs': 0.011303} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.154904] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.155174] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.155450] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.155603] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.155783] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.156106] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.156569] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.156825] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dad4e327-64cb-4a80-8deb-309effa9262e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.159561] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94c50d44-b748-4b17-80f9-89958dbac267 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.165390] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 859.165390] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52626316-1261-a949-245c-b4808eab5b54" [ 859.165390] env[63197]: _type = "Task" [ 859.165390] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.172459] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.172688] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 859.174154] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-739a134f-0742-41f2-906d-33dbb03e1097 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.179656] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52626316-1261-a949-245c-b4808eab5b54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.180626] env[63197]: DEBUG nova.network.neutron [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Updated VIF entry in instance network info cache for port 375c2467-68af-4a0b-8cce-2209b198b1d8. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.181022] env[63197]: DEBUG nova.network.neutron [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Updating instance_info_cache with network_info: [{"id": "375c2467-68af-4a0b-8cce-2209b198b1d8", "address": "fa:16:3e:1f:85:55", "network": {"id": "ae948c4d-6288-4540-b9e5-62f26a6719ae", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1174851250-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "185f0eaa93b5435295800f0c7642cb4e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap375c2467-68", "ovs_interfaceid": "375c2467-68af-4a0b-8cce-2209b198b1d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.188232] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 859.188232] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f0d102-d49b-8550-643e-486abeb3bd64" [ 859.188232] env[63197]: _type = "Task" [ 859.188232] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.202256] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f0d102-d49b-8550-643e-486abeb3bd64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.259093] env[63197]: DEBUG nova.network.neutron [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.436187] env[63197]: DEBUG nova.scheduler.client.report [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.467280] env[63197]: DEBUG oslo_vmware.api [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364108, 'name': PowerOnVM_Task, 'duration_secs': 0.799556} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.467664] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.467934] env[63197]: INFO nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Took 7.86 seconds to spawn the instance on the hypervisor. [ 859.468181] env[63197]: DEBUG nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.468991] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b402cb9-97fd-4ae6-a0be-a46ab8c6c68f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.677900] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52626316-1261-a949-245c-b4808eab5b54, 'name': SearchDatastore_Task, 'duration_secs': 0.020426} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.678471] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.678749] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.678983] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.683541] env[63197]: DEBUG oslo_concurrency.lockutils [req-9e26c0e6-8d6c-4390-a4cd-9a3739b7d13a req-d199ad46-bf86-4435-9925-60a1add12404 service nova] Releasing lock "refresh_cache-c55b5b2b-5ecc-43bb-a279-7370cd9ac722" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.697876] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f0d102-d49b-8550-643e-486abeb3bd64, 'name': SearchDatastore_Task, 'duration_secs': 0.016784} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.698629] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92dd74d2-f3e0-4682-9177-c8c137aaa3a7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.703904] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 859.703904] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ebfc13-ca3d-ebdb-b352-da6c000abba6" [ 859.703904] env[63197]: _type = "Task" [ 859.703904] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.711819] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ebfc13-ca3d-ebdb-b352-da6c000abba6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.763622] env[63197]: INFO nova.compute.manager [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] [instance: fc4d4223-b0cb-4a40-bd5f-62ec740b57d3] Took 1.03 seconds to deallocate network for instance. [ 859.943023] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.850s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.943023] env[63197]: ERROR nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. [ 859.943023] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Traceback (most recent call last): [ 859.943023] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 859.943023] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self.driver.spawn(context, instance, image_meta, [ 859.943023] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 859.943023] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 859.943023] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 859.943023] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] vm_ref = self.build_virtual_machine(instance, [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] vif_infos = vmwarevif.get_vif_info(self._session, [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] for vif in network_info: [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] return self._sync_wrapper(fn, *args, **kwargs) [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self.wait() [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self[:] = self._gt.wait() [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] return self._exit_event.wait() [ 859.943524] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] current.throw(*self._exc) [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] result = function(*args, **kwargs) [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] return func(*args, **kwargs) [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] raise e [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] nwinfo = self.network_api.allocate_for_instance( [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] created_port_ids = self._update_ports_for_instance( [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 859.943864] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] with excutils.save_and_reraise_exception(): [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] self.force_reraise() [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] raise self.value [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] updated_port = self._update_port( [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] _ensure_no_port_binding_failure(port) [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] raise exception.PortBindingFailed(port_id=port['id']) [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] nova.exception.PortBindingFailed: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. [ 859.944209] env[63197]: ERROR nova.compute.manager [instance: c005d088-ee27-4d20-9db4-65b47227f22a] [ 859.944477] env[63197]: DEBUG nova.compute.utils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 859.948017] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.971s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.948793] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Build of instance c005d088-ee27-4d20-9db4-65b47227f22a was re-scheduled: Binding failed for port cd260090-5f70-4fb0-8355-d52e82eb44f9, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 859.949419] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 859.950021] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquiring lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.950021] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Acquired lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.950372] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.988942] env[63197]: INFO nova.compute.manager [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Took 36.48 seconds to build instance. [ 860.214204] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ebfc13-ca3d-ebdb-b352-da6c000abba6, 'name': SearchDatastore_Task, 'duration_secs': 0.009242} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.214563] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.214887] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] c55b5b2b-5ecc-43bb-a279-7370cd9ac722/c55b5b2b-5ecc-43bb-a279-7370cd9ac722.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.215251] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.215522] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.215783] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e456b88-618d-4fb7-81e8-e5103aa8e695 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.217885] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb137dd7-be69-4988-9c48-9089cdbdb6ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.225157] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 860.225157] env[63197]: value = "task-1364109" [ 860.225157] env[63197]: _type = "Task" [ 860.225157] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.229423] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.230049] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.231015] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d804158f-a55d-44b1-9fba-4bb273316e8b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.237845] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364109, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.241814] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 860.241814] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522abb46-0d0d-0087-8d1d-e159a22ca5bf" [ 860.241814] env[63197]: _type = "Task" [ 860.241814] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.253398] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522abb46-0d0d-0087-8d1d-e159a22ca5bf, 'name': SearchDatastore_Task, 'duration_secs': 0.007935} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.253505] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d11b86a-6763-4b22-b76e-470adb9e0373 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.259165] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 860.259165] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524fa51e-0f16-278b-fa33-9a488e2245c1" [ 860.259165] env[63197]: _type = "Task" [ 860.259165] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.272679] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524fa51e-0f16-278b-fa33-9a488e2245c1, 'name': SearchDatastore_Task, 'duration_secs': 0.007467} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.273262] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.273619] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.273953] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3fb958c3-eba1-4945-8f95-3caf3e073dc0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.280910] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 860.280910] env[63197]: value = "task-1364110" [ 860.280910] env[63197]: _type = "Task" [ 860.280910] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.288889] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364110, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.481188] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.490540] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba152f05-4770-4b45-a861-33cd20eb67bc tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.838s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.622174] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.739369] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364109, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475538} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.740430] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] c55b5b2b-5ecc-43bb-a279-7370cd9ac722/c55b5b2b-5ecc-43bb-a279-7370cd9ac722.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.740430] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.740430] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f338356-cb1a-4172-9a27-13cc56b0198b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.747966] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 860.747966] env[63197]: value = "task-1364111" [ 860.747966] env[63197]: _type = "Task" [ 860.747966] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.756337] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364111, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.773306] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0af291d-bae3-4de5-992f-42074d99eee9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.788596] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd62d7a-cc54-43ca-8156-c04805678815 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.799835] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364110, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.831834] env[63197]: INFO nova.scheduler.client.report [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Deleted allocations for instance fc4d4223-b0cb-4a40-bd5f-62ec740b57d3 [ 860.840011] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea7a8c0-d44d-4c1b-b29c-948daacd7815 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.851823] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91e925f-4aee-4776-a930-322fb6a078fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.868935] env[63197]: DEBUG nova.compute.provider_tree [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.993913] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.126503] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Releasing lock "refresh_cache-c005d088-ee27-4d20-9db4-65b47227f22a" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.126748] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 861.126932] env[63197]: DEBUG nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.127113] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 861.143623] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.262068] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364111, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084891} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.262397] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.263226] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab744cde-0cc1-4e7c-af49-1359f853b46b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.290504] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] c55b5b2b-5ecc-43bb-a279-7370cd9ac722/c55b5b2b-5ecc-43bb-a279-7370cd9ac722.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.291250] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49d17ee3-db0d-447c-a105-12b009435955 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.317277] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364110, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738855} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.319263] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 861.319476] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.319777] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 861.319777] env[63197]: value = "task-1364112" [ 861.319777] env[63197]: _type = "Task" [ 861.319777] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.320601] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d9c7af4-d92a-4402-800c-d35ab25f7cb5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.332118] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364112, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.333667] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 861.333667] env[63197]: value = "task-1364113" [ 861.333667] env[63197]: _type = "Task" [ 861.333667] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.341233] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.347705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-28a171b8-e0b9-4a81-8bcf-4d851fc11ced tempest-ServerTagsTestJSON-1754529947 tempest-ServerTagsTestJSON-1754529947-project-member] Lock "fc4d4223-b0cb-4a40-bd5f-62ec740b57d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.804s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.375734] env[63197]: DEBUG nova.scheduler.client.report [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.526978] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.647378] env[63197]: DEBUG nova.network.neutron [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.833126] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364112, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.841990] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.224731} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.842385] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.843265] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aee4e64-a5fc-4bf4-98b0-78241ebdac8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.849222] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.880123] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.880708] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.881221] env[63197]: ERROR nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Traceback (most recent call last): [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self.driver.spawn(context, instance, image_meta, [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] vm_ref = self.build_virtual_machine(instance, [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 861.881221] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] for vif in network_info: [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] return self._sync_wrapper(fn, *args, **kwargs) [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self.wait() [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self[:] = self._gt.wait() [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] return self._exit_event.wait() [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] current.throw(*self._exc) [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 861.881539] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] result = function(*args, **kwargs) [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] return func(*args, **kwargs) [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] raise e [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] nwinfo = self.network_api.allocate_for_instance( [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] created_port_ids = self._update_ports_for_instance( [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] with excutils.save_and_reraise_exception(): [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] self.force_reraise() [ 861.881914] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] raise self.value [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] updated_port = self._update_port( [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] _ensure_no_port_binding_failure(port) [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] raise exception.PortBindingFailed(port_id=port['id']) [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] nova.exception.PortBindingFailed: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. [ 861.882268] env[63197]: ERROR nova.compute.manager [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] [ 861.882268] env[63197]: DEBUG nova.compute.utils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 861.883300] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c68015c-a84f-4628-b88c-3fcf6a8813f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.899043] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Build of instance 643622a1-cb00-4829-a83a-a347e6e50e5d was re-scheduled: Binding failed for port 7455e2c7-9551-4a3b-b203-818e965b59ad, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 861.899043] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 861.899043] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.899043] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.899300] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.900358] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.948s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.902136] env[63197]: INFO nova.compute.claims [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.911522] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 861.911522] env[63197]: value = "task-1364114" [ 861.911522] env[63197]: _type = "Task" [ 861.911522] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.919605] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364114, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.063730] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "232ea430-6a11-4dbc-9da9-0f734471d03b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.064073] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.151649] env[63197]: INFO nova.compute.manager [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] [instance: c005d088-ee27-4d20-9db4-65b47227f22a] Took 1.02 seconds to deallocate network for instance. [ 862.333875] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364112, 'name': ReconfigVM_Task, 'duration_secs': 0.718906} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.334374] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Reconfigured VM instance instance-0000003d to attach disk [datastore2] c55b5b2b-5ecc-43bb-a279-7370cd9ac722/c55b5b2b-5ecc-43bb-a279-7370cd9ac722.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.335255] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f26fa294-3b55-4905-b9f4-69a878820890 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.341885] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 862.341885] env[63197]: value = "task-1364115" [ 862.341885] env[63197]: _type = "Task" [ 862.341885] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.350455] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364115, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.385415] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.419306] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364114, 'name': ReconfigVM_Task, 'duration_secs': 0.489245} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.420217] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.420804] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae2a3ec2-d71c-4178-b2b2-591e1d117c4e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.426470] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 862.426470] env[63197]: value = "task-1364116" [ 862.426470] env[63197]: _type = "Task" [ 862.426470] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.435164] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364116, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.487440] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.652838] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.852081] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364115, 'name': Rename_Task, 'duration_secs': 0.391601} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.852408] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.852685] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ef03e4d-ee87-44f1-b82a-21663e104feb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.859469] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 862.859469] env[63197]: value = "task-1364117" [ 862.859469] env[63197]: _type = "Task" [ 862.859469] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.867919] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.868210] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.872231] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364117, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.936032] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364116, 'name': Rename_Task, 'duration_secs': 0.348127} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.936520] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.936906] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bed5d21-2bf0-43ad-94af-515022020c0f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.945625] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 862.945625] env[63197]: value = "task-1364118" [ 862.945625] env[63197]: _type = "Task" [ 862.945625] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.953690] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364118, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.142875] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d17bb2-cde9-4c73-af89-ebba6081b620 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.150097] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181ad415-9756-4e30-b342-c77fa58e94a7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.155690] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "refresh_cache-643622a1-cb00-4829-a83a-a347e6e50e5d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.155932] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 863.156213] env[63197]: DEBUG nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.156397] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 863.186281] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d374569-4c8c-4f74-bdae-bbacf0049c0f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.192716] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7575ad71-e21a-463b-80ba-7f7c7a334575 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.206101] env[63197]: DEBUG nova.compute.provider_tree [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.208209] env[63197]: INFO nova.scheduler.client.report [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Deleted allocations for instance c005d088-ee27-4d20-9db4-65b47227f22a [ 863.214028] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.371823] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364117, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.455185] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364118, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.717324] env[63197]: DEBUG nova.scheduler.client.report [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.721980] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2cf4540f-317e-4b0b-aa51-34350467c787 tempest-VolumesAdminNegativeTest-23941391 tempest-VolumesAdminNegativeTest-23941391-project-member] Lock "c005d088-ee27-4d20-9db4-65b47227f22a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.635s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.722307] env[63197]: DEBUG nova.network.neutron [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.873107] env[63197]: DEBUG oslo_vmware.api [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364117, 'name': PowerOnVM_Task, 'duration_secs': 0.813847} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.873107] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.873107] env[63197]: INFO nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Took 8.07 seconds to spawn the instance on the hypervisor. [ 863.873107] env[63197]: DEBUG nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.873107] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88755a96-6ddc-437d-af36-252e7237a82b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.957538] env[63197]: DEBUG oslo_vmware.api [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364118, 'name': PowerOnVM_Task, 'duration_secs': 0.699768} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.957842] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.958036] env[63197]: INFO nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Took 5.69 seconds to spawn the instance on the hypervisor. [ 863.958222] env[63197]: DEBUG nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.959098] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e985e2-91e5-4349-a5e3-e939438d3a49 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.225948] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.226513] env[63197]: DEBUG nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.230192] env[63197]: INFO nova.compute.manager [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 643622a1-cb00-4829-a83a-a347e6e50e5d] Took 1.07 seconds to deallocate network for instance. [ 864.232886] env[63197]: DEBUG nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 864.236798] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.814s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.393349] env[63197]: INFO nova.compute.manager [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Took 27.34 seconds to build instance. [ 864.474391] env[63197]: INFO nova.compute.manager [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Took 25.62 seconds to build instance. [ 864.732545] env[63197]: DEBUG nova.compute.utils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.734021] env[63197]: DEBUG nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.734221] env[63197]: DEBUG nova.network.neutron [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 864.767451] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.816402] env[63197]: DEBUG nova.policy [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.894799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f6840347-ab39-4294-9ba1-a322d90bd697 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.162s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.976975] env[63197]: DEBUG oslo_concurrency.lockutils [None req-82d47392-146f-4770-ad51-6e42eb3283cc tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.352s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.994761] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68aea9a2-ff30-4429-85dd-048b5c307840 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.003011] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302e2e89-562e-482b-9ffd-1d9bf603814f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.036015] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836f9fb2-5013-4cd7-935c-8c276578b145 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.043424] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a093497f-393a-4299-965d-7c7190299612 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.056843] env[63197]: DEBUG nova.compute.provider_tree [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.241812] env[63197]: DEBUG nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.271895] env[63197]: INFO nova.scheduler.client.report [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Deleted allocations for instance 643622a1-cb00-4829-a83a-a347e6e50e5d [ 865.370616] env[63197]: DEBUG nova.network.neutron [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Successfully created port: 1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.397199] env[63197]: DEBUG nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 865.479811] env[63197]: DEBUG nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 865.560543] env[63197]: DEBUG nova.scheduler.client.report [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.793027] env[63197]: INFO nova.compute.manager [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Rebuilding instance [ 865.793027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-424ef2a8-2100-4ccf-bc02-0a0834b4994e tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "643622a1-cb00-4829-a83a-a347e6e50e5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.669s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.848279] env[63197]: DEBUG nova.compute.manager [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 865.851044] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf5be48-7bd7-4ef1-bb0e-cb473335e3e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.923396] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.005744] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.065538] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.830s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.067566] env[63197]: ERROR nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Traceback (most recent call last): [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self.driver.spawn(context, instance, image_meta, [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] vm_ref = self.build_virtual_machine(instance, [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 866.067566] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] for vif in network_info: [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] return self._sync_wrapper(fn, *args, **kwargs) [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self.wait() [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self[:] = self._gt.wait() [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] return self._exit_event.wait() [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] current.throw(*self._exc) [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 866.067896] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] result = function(*args, **kwargs) [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] return func(*args, **kwargs) [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] raise e [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] nwinfo = self.network_api.allocate_for_instance( [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] created_port_ids = self._update_ports_for_instance( [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] with excutils.save_and_reraise_exception(): [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] self.force_reraise() [ 866.068215] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] raise self.value [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] updated_port = self._update_port( [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] _ensure_no_port_binding_failure(port) [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] raise exception.PortBindingFailed(port_id=port['id']) [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] nova.exception.PortBindingFailed: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. [ 866.068510] env[63197]: ERROR nova.compute.manager [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] [ 866.068510] env[63197]: DEBUG nova.compute.utils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 866.071455] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Build of instance 92ee8e7a-c169-44d8-8f74-84590b4f62e9 was re-scheduled: Binding failed for port a31e966e-24dd-4793-8bb1-fe856caf9242, please check neutron logs for more information. {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 866.071932] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Unplugging VIFs for instance {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 866.072194] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.072349] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.072510] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.076383] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.040s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.076383] env[63197]: INFO nova.compute.claims [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.099177] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.218484] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.252572] env[63197]: DEBUG nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.275192] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.275478] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.275600] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.275773] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.275984] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.276093] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.276274] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.276424] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.276584] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.276733] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.276892] env[63197]: DEBUG nova.virt.hardware [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.277788] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de79224d-255b-4883-88a9-bdeae1e06a92 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.286215] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d93761-1ed0-41d8-88ff-976f7b8d311e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.299707] env[63197]: DEBUG nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.360950] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.361781] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f5684b5-ec69-4581-9ef6-4e0872fa97d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.368339] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 866.368339] env[63197]: value = "task-1364119" [ 866.368339] env[63197]: _type = "Task" [ 866.368339] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.372048] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.372292] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.372487] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.372665] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.372834] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.375019] env[63197]: INFO nova.compute.manager [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Terminating instance [ 866.376865] env[63197]: DEBUG nova.compute.manager [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.377106] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.380908] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38bf7ed2-a43e-4e49-9f99-60d9d58db024 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.383682] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364119, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.390204] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.390495] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b1c4545-f209-43cc-b138-90835afb5b52 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.396368] env[63197]: DEBUG oslo_vmware.api [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 866.396368] env[63197]: value = "task-1364120" [ 866.396368] env[63197]: _type = "Task" [ 866.396368] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.408236] env[63197]: DEBUG oslo_vmware.api [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.721609] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-92ee8e7a-c169-44d8-8f74-84590b4f62e9" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.721897] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63197) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 866.722098] env[63197]: DEBUG nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 866.722276] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 866.763431] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.835079] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.878480] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364119, 'name': PowerOffVM_Task, 'duration_secs': 0.203473} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.878751] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.878955] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.879767] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dfba0d-bcbc-4f9b-b220-983524bfeadb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.886794] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.887045] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e789a262-4d1c-4bdf-9eb8-a53b59142021 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.906593] env[63197]: DEBUG oslo_vmware.api [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364120, 'name': PowerOffVM_Task, 'duration_secs': 0.191209} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.906860] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.907041] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.907311] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3354bee2-8b6c-4a54-80b2-e55fb15691ec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.911998] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.912220] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.912518] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Deleting the datastore file [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.912652] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95af1a97-a8a5-40f0-b1a6-84fd24ae718a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.919394] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 866.919394] env[63197]: value = "task-1364123" [ 866.919394] env[63197]: _type = "Task" [ 866.919394] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.927264] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364123, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.989430] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.989807] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.989936] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Deleting the datastore file [datastore2] c55b5b2b-5ecc-43bb-a279-7370cd9ac722 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.990238] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88b02744-34dd-4fc8-a975-7882322ace33 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.998443] env[63197]: DEBUG oslo_vmware.api [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for the task: (returnval){ [ 866.998443] env[63197]: value = "task-1364124" [ 866.998443] env[63197]: _type = "Task" [ 866.998443] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.005504] env[63197]: DEBUG oslo_vmware.api [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.266416] env[63197]: DEBUG nova.network.neutron [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.288801] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d28baf-f895-4473-b49d-6e3e0e7bd277 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.297304] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d236b913-a66c-4e93-8c42-6765fc7eeb8c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.328166] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba19f0f-78e1-4630-87a0-0a9792118cfb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.335354] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2cbcdc-86ea-4ed6-9909-ebc9c86021ca {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.353433] env[63197]: DEBUG nova.compute.provider_tree [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.429750] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364123, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.115178} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.430014] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.430205] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.430412] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.501609] env[63197]: DEBUG nova.compute.manager [req-df2d5175-5ace-44d9-bb55-2eb569089724 req-6fbe9448-891e-426b-90cd-9c4723d00051 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-vif-plugged-1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.501609] env[63197]: DEBUG oslo_concurrency.lockutils [req-df2d5175-5ace-44d9-bb55-2eb569089724 req-6fbe9448-891e-426b-90cd-9c4723d00051 service nova] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.501609] env[63197]: DEBUG oslo_concurrency.lockutils [req-df2d5175-5ace-44d9-bb55-2eb569089724 req-6fbe9448-891e-426b-90cd-9c4723d00051 service nova] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.501609] env[63197]: DEBUG oslo_concurrency.lockutils [req-df2d5175-5ace-44d9-bb55-2eb569089724 req-6fbe9448-891e-426b-90cd-9c4723d00051 service nova] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.501609] env[63197]: DEBUG nova.compute.manager [req-df2d5175-5ace-44d9-bb55-2eb569089724 req-6fbe9448-891e-426b-90cd-9c4723d00051 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] No waiting events found dispatching network-vif-plugged-1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 867.501848] env[63197]: WARNING nova.compute.manager [req-df2d5175-5ace-44d9-bb55-2eb569089724 req-6fbe9448-891e-426b-90cd-9c4723d00051 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received unexpected event network-vif-plugged-1d3f4b96-07ba-4875-9441-753c3b9855a9 for instance with vm_state building and task_state spawning. [ 867.512035] env[63197]: DEBUG oslo_vmware.api [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Task: {'id': task-1364124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140556} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.512035] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.512035] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.512035] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.512550] env[63197]: INFO nova.compute.manager [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Took 1.14 seconds to destroy the instance on the hypervisor. [ 867.512550] env[63197]: DEBUG oslo.service.loopingcall [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.512550] env[63197]: DEBUG nova.compute.manager [-] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.512701] env[63197]: DEBUG nova.network.neutron [-] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 867.601781] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "9d8c131d-1183-4508-ae2d-28e38a50e58d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.602208] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.720304] env[63197]: DEBUG nova.network.neutron [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Successfully updated port: 1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.772665] env[63197]: INFO nova.compute.manager [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 92ee8e7a-c169-44d8-8f74-84590b4f62e9] Took 1.05 seconds to deallocate network for instance. [ 867.777582] env[63197]: DEBUG nova.compute.manager [req-7d8a6161-751e-4cf0-a65b-192ad4163ca6 req-b8cadf7e-3259-4346-a2eb-37cae1d2bdca service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Received event network-vif-deleted-375c2467-68af-4a0b-8cce-2209b198b1d8 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.777878] env[63197]: INFO nova.compute.manager [req-7d8a6161-751e-4cf0-a65b-192ad4163ca6 req-b8cadf7e-3259-4346-a2eb-37cae1d2bdca service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Neutron deleted interface 375c2467-68af-4a0b-8cce-2209b198b1d8; detaching it from the instance and deleting it from the info cache [ 867.778109] env[63197]: DEBUG nova.network.neutron [req-7d8a6161-751e-4cf0-a65b-192ad4163ca6 req-b8cadf7e-3259-4346-a2eb-37cae1d2bdca service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.855241] env[63197]: DEBUG nova.scheduler.client.report [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.224733] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.224902] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.225056] env[63197]: DEBUG nova.network.neutron [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.251621] env[63197]: DEBUG nova.network.neutron [-] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.285198] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42e4ab97-03d8-41d4-b681-b7b2146001b2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.301838] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19e695b5-61d0-438e-857b-f41ba5939707 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.320356] env[63197]: DEBUG nova.compute.manager [req-7d8a6161-751e-4cf0-a65b-192ad4163ca6 req-b8cadf7e-3259-4346-a2eb-37cae1d2bdca service nova] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Detach interface failed, port_id=375c2467-68af-4a0b-8cce-2209b198b1d8, reason: Instance c55b5b2b-5ecc-43bb-a279-7370cd9ac722 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 868.360017] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.360403] env[63197]: DEBUG nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.363036] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.145s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.364447] env[63197]: INFO nova.compute.claims [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.464850] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.465122] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.465281] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.465459] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.465603] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.465749] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.465952] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.466156] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.466367] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.466525] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.466700] env[63197]: DEBUG nova.virt.hardware [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.467863] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350e7e46-b111-4f91-89f5-a93dcc46a943 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.476101] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d58cff8-f06d-430c-8a8e-d3b153c20225 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.489761] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 868.495538] env[63197]: DEBUG oslo.service.loopingcall [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.495801] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 868.496015] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-603d967b-63c9-49fe-b700-dc0c45c5ca7a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.513571] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 868.513571] env[63197]: value = "task-1364125" [ 868.513571] env[63197]: _type = "Task" [ 868.513571] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.520880] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364125, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.754297] env[63197]: INFO nova.compute.manager [-] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Took 1.24 seconds to deallocate network for instance. [ 868.777148] env[63197]: DEBUG nova.network.neutron [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.817564] env[63197]: INFO nova.scheduler.client.report [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted allocations for instance 92ee8e7a-c169-44d8-8f74-84590b4f62e9 [ 868.869362] env[63197]: DEBUG nova.compute.utils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.873574] env[63197]: DEBUG nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.873752] env[63197]: DEBUG nova.network.neutron [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.919929] env[63197]: DEBUG nova.policy [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '641264b91b9340f393841313bd05ea4d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b568f78ebda4dbaafefa681900293e7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.960574] env[63197]: DEBUG nova.network.neutron [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.023505] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364125, 'name': CreateVM_Task, 'duration_secs': 0.297713} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.023862] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 869.024140] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.024310] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.024704] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 869.024879] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f78bc89-d647-41b2-922c-942f36a3a7fd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.029286] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 869.029286] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527e1f2d-71f9-1d09-fc9e-1536a87dd3f7" [ 869.029286] env[63197]: _type = "Task" [ 869.029286] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.037069] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527e1f2d-71f9-1d09-fc9e-1536a87dd3f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.197687] env[63197]: DEBUG nova.network.neutron [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Successfully created port: ca0a6156-1f24-4a5c-867d-f35a33c87876 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.264908] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.329635] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7643a66b-cf44-4c5e-b892-2b1a678b368c tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "92ee8e7a-c169-44d8-8f74-84590b4f62e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.720s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.380634] env[63197]: DEBUG nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.465219] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.465532] env[63197]: DEBUG nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Instance network_info: |[{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 869.466781] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:3f:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d3f4b96-07ba-4875-9441-753c3b9855a9', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.474251] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Creating folder: Project (d8db3cc951174f6192ff954ff4d704de). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.474516] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bce5315-6851-49a3-84fe-9326f21ff1f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.489132] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Created folder: Project (d8db3cc951174f6192ff954ff4d704de) in parent group-v290286. [ 869.490784] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Creating folder: Instances. Parent ref: group-v290316. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 869.493877] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2e4d9959-b61b-47c8-b85a-711ada2ec5fe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.503025] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Created folder: Instances in parent group-v290316. [ 869.503166] env[63197]: DEBUG oslo.service.loopingcall [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.503256] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.503450] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80a447fa-e1a0-4452-b42d-7da0743c479e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.532088] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 869.532088] env[63197]: value = "task-1364128" [ 869.532088] env[63197]: _type = "Task" [ 869.532088] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.547798] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527e1f2d-71f9-1d09-fc9e-1536a87dd3f7, 'name': SearchDatastore_Task, 'duration_secs': 0.009025} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.551074] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.551347] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.551703] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.552157] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.552157] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.552157] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364128, 'name': CreateVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.552549] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-61e761dc-12d1-49aa-98f4-2d46c72bb17e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.564271] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.564457] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 869.565437] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b88f2f0c-b618-4f01-862f-1c5326bf8c21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.573695] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 869.573695] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c524d3-e3bc-5d0d-40b2-e97c86fa4e1e" [ 869.573695] env[63197]: _type = "Task" [ 869.573695] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.583397] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c524d3-e3bc-5d0d-40b2-e97c86fa4e1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.606406] env[63197]: DEBUG nova.compute.manager [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-changed-1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.606599] env[63197]: DEBUG nova.compute.manager [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing instance network info cache due to event network-changed-1d3f4b96-07ba-4875-9441-753c3b9855a9. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.606837] env[63197]: DEBUG oslo_concurrency.lockutils [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.606938] env[63197]: DEBUG oslo_concurrency.lockutils [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.607149] env[63197]: DEBUG nova.network.neutron [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing network info cache for port 1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 869.642096] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c5217c-52cf-4917-887e-99f3c626bac1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.649959] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825d5e5e-c117-40f9-b93b-9234ed57723f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.683922] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c832453c-0d6c-45ca-a7e1-b4fce105794e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.691999] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92955b57-9e81-462b-8182-8974edfcda2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.706019] env[63197]: DEBUG nova.compute.provider_tree [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.836189] env[63197]: DEBUG nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 870.046168] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364128, 'name': CreateVM_Task, 'duration_secs': 0.391405} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.046544] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.047396] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.047579] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.048007] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.048194] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a22ba798-c5e6-4dbd-909b-aaaca58915cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.053103] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 870.053103] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d0db52-9c92-537f-ca08-8fca9c85bcb9" [ 870.053103] env[63197]: _type = "Task" [ 870.053103] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.061577] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d0db52-9c92-537f-ca08-8fca9c85bcb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.083361] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c524d3-e3bc-5d0d-40b2-e97c86fa4e1e, 'name': SearchDatastore_Task, 'duration_secs': 0.009108} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.084149] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d5f9308-404b-432b-9b4c-24690d2be8c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.089176] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 870.089176] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527d4827-243a-e7a6-6b3c-61dd8979db26" [ 870.089176] env[63197]: _type = "Task" [ 870.089176] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.098250] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527d4827-243a-e7a6-6b3c-61dd8979db26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.209627] env[63197]: DEBUG nova.scheduler.client.report [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.357623] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.394358] env[63197]: DEBUG nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.416536] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.416756] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.416904] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.417096] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.417238] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.417378] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.417575] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.417737] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.417902] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.418068] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.418236] env[63197]: DEBUG nova.virt.hardware [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.421544] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5cd79c0-6d71-4eb7-adc7-c2e753982e87 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.427411] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476faff4-3da2-4224-a010-fc8004c3d5d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.475098] env[63197]: DEBUG nova.network.neutron [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updated VIF entry in instance network info cache for port 1d3f4b96-07ba-4875-9441-753c3b9855a9. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.475457] env[63197]: DEBUG nova.network.neutron [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.563643] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d0db52-9c92-537f-ca08-8fca9c85bcb9, 'name': SearchDatastore_Task, 'duration_secs': 0.009791} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.563980] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.564235] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 870.564467] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.564612] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.564784] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 870.565056] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e4fb75c-cca8-4888-9266-879c924929a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.573163] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 870.573347] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 870.574087] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47efd0bb-a479-4c97-a668-f48bd07a29e0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.579315] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 870.579315] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d53be9-d231-b34d-3b08-9bcbbcc0ab51" [ 870.579315] env[63197]: _type = "Task" [ 870.579315] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.586854] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d53be9-d231-b34d-3b08-9bcbbcc0ab51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.599013] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527d4827-243a-e7a6-6b3c-61dd8979db26, 'name': SearchDatastore_Task, 'duration_secs': 0.008448} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.603074] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.603074] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 870.603074] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3300f9c8-44b5-44c7-ba3d-c677033f77b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.607175] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 870.607175] env[63197]: value = "task-1364129" [ 870.607175] env[63197]: _type = "Task" [ 870.607175] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.615088] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.721868] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.722694] env[63197]: DEBUG nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 870.731257] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.204s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.733831] env[63197]: INFO nova.compute.claims [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.738824] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "a1f365de-b634-481a-b2e8-9bfc801a47f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.739359] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.978629] env[63197]: DEBUG oslo_concurrency.lockutils [req-a3eb207e-37a3-4ee9-8def-90a3a6e53cd4 req-d25e7921-6c00-407c-923b-b8f5bf5499e7 service nova] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.990540] env[63197]: DEBUG nova.network.neutron [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Successfully updated port: ca0a6156-1f24-4a5c-867d-f35a33c87876 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.094077] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d53be9-d231-b34d-3b08-9bcbbcc0ab51, 'name': SearchDatastore_Task, 'duration_secs': 0.00961} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.094077] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c8d2e43-39af-4ee8-a72b-b162ef9cda70 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.098211] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 871.098211] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528af15a-9775-affd-f1ae-fae9a702d0fe" [ 871.098211] env[63197]: _type = "Task" [ 871.098211] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.108316] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528af15a-9775-affd-f1ae-fae9a702d0fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.117872] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364129, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.234256] env[63197]: DEBUG nova.compute.utils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.234823] env[63197]: DEBUG nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.236360] env[63197]: DEBUG nova.network.neutron [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 871.277809] env[63197]: DEBUG nova.policy [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '17a795e0ae3f44ac94e8bdd623a8abda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '153c4635bb6740308040698319aff9be', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.492884] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "refresh_cache-72f9b0c4-69b9-49f2-8665-ff77151883af" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.493179] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquired lock "refresh_cache-72f9b0c4-69b9-49f2-8665-ff77151883af" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.493256] env[63197]: DEBUG nova.network.neutron [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.572354] env[63197]: DEBUG nova.network.neutron [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Successfully created port: eae8fdec-5c12-41cd-b92e-8673e5d68cb0 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.615987] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528af15a-9775-affd-f1ae-fae9a702d0fe, 'name': SearchDatastore_Task, 'duration_secs': 0.022768} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.618893] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.619494] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 9c1caa2b-c369-425a-8726-cddadf06f338/9c1caa2b-c369-425a-8726-cddadf06f338.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.619494] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29cee9d7-cf56-4c24-bfb0-3c246adadf6d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.624599] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364129, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.55009} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.625705] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 871.625871] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 871.626883] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17ecf253-4c49-423a-ae1b-df820f62169b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.629671] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 871.629671] env[63197]: value = "task-1364130" [ 871.629671] env[63197]: _type = "Task" [ 871.629671] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.634474] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 871.634474] env[63197]: value = "task-1364131" [ 871.634474] env[63197]: _type = "Task" [ 871.634474] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.642699] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.651596] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364131, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.692779] env[63197]: DEBUG nova.compute.manager [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Received event network-vif-plugged-ca0a6156-1f24-4a5c-867d-f35a33c87876 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.692779] env[63197]: DEBUG oslo_concurrency.lockutils [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] Acquiring lock "72f9b0c4-69b9-49f2-8665-ff77151883af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.692779] env[63197]: DEBUG oslo_concurrency.lockutils [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.692779] env[63197]: DEBUG oslo_concurrency.lockutils [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.694092] env[63197]: DEBUG nova.compute.manager [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] No waiting events found dispatching network-vif-plugged-ca0a6156-1f24-4a5c-867d-f35a33c87876 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.694457] env[63197]: WARNING nova.compute.manager [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Received unexpected event network-vif-plugged-ca0a6156-1f24-4a5c-867d-f35a33c87876 for instance with vm_state building and task_state spawning. [ 871.694763] env[63197]: DEBUG nova.compute.manager [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Received event network-changed-ca0a6156-1f24-4a5c-867d-f35a33c87876 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.695048] env[63197]: DEBUG nova.compute.manager [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Refreshing instance network info cache due to event network-changed-ca0a6156-1f24-4a5c-867d-f35a33c87876. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.695366] env[63197]: DEBUG oslo_concurrency.lockutils [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] Acquiring lock "refresh_cache-72f9b0c4-69b9-49f2-8665-ff77151883af" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.740388] env[63197]: DEBUG nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.053780] env[63197]: DEBUG nova.network.neutron [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.057710] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b0d379-65ec-4c3f-b9ca-31af03de161a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.066983] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a65825-8979-47cb-84bd-fe112b23298f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.107681] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4702fdb4-3a56-4f48-97b2-17c03e53882d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.115058] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1865db8-47a8-47c8-817b-11b6e1f73f08 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.129936] env[63197]: DEBUG nova.compute.provider_tree [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.145323] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364130, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496056} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.148201] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 9c1caa2b-c369-425a-8726-cddadf06f338/9c1caa2b-c369-425a-8726-cddadf06f338.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.148460] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.148965] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-138549f2-7dbc-4f58-85c7-f659ff7f4f80 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.155404] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364131, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066054} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.155404] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.155976] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c572bd-e7be-4b68-a511-53b8e1cef3af {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.159420] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 872.159420] env[63197]: value = "task-1364132" [ 872.159420] env[63197]: _type = "Task" [ 872.159420] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.178216] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.178959] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0105669d-bdcd-4f74-b7d0-b1c0d366a5ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.199193] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364132, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.204669] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 872.204669] env[63197]: value = "task-1364133" [ 872.204669] env[63197]: _type = "Task" [ 872.204669] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.213508] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.354301] env[63197]: DEBUG nova.network.neutron [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Updating instance_info_cache with network_info: [{"id": "ca0a6156-1f24-4a5c-867d-f35a33c87876", "address": "fa:16:3e:a9:4a:d2", "network": {"id": "e437d47b-36c0-481d-acb7-fc82b08ab41e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1625379393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b568f78ebda4dbaafefa681900293e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca0a6156-1f", "ovs_interfaceid": "ca0a6156-1f24-4a5c-867d-f35a33c87876", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.635717] env[63197]: DEBUG nova.scheduler.client.report [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.672345] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364132, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.217809} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.672630] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 872.673558] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4164812d-89fc-40a9-95e5-05fe3829905e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.697225] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 9c1caa2b-c369-425a-8726-cddadf06f338/9c1caa2b-c369-425a-8726-cddadf06f338.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 872.698016] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c3d3eec-72f7-4695-b654-473e3848206b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.724630] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364133, 'name': ReconfigVM_Task, 'duration_secs': 0.306265} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.726334] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e/3b7aa0b0-d115-4026-a984-0e9fc73f2d2e.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.727422] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 872.727422] env[63197]: value = "task-1364134" [ 872.727422] env[63197]: _type = "Task" [ 872.727422] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.727585] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-17f0f976-125a-48a4-93eb-035a54ca616d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.737563] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364134, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.738777] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 872.738777] env[63197]: value = "task-1364135" [ 872.738777] env[63197]: _type = "Task" [ 872.738777] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.749761] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364135, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.753085] env[63197]: DEBUG nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 872.785057] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.785057] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.785057] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.785272] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.785272] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.785539] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.785884] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.786210] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.786522] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.786826] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.787158] env[63197]: DEBUG nova.virt.hardware [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.788234] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a2e4ed-df64-4693-a629-6a2ee8857f50 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.797691] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169d72a1-2997-467b-bb0b-bd1fd8b72401 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.860022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Releasing lock "refresh_cache-72f9b0c4-69b9-49f2-8665-ff77151883af" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.860022] env[63197]: DEBUG nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Instance network_info: |[{"id": "ca0a6156-1f24-4a5c-867d-f35a33c87876", "address": "fa:16:3e:a9:4a:d2", "network": {"id": "e437d47b-36c0-481d-acb7-fc82b08ab41e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1625379393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b568f78ebda4dbaafefa681900293e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca0a6156-1f", "ovs_interfaceid": "ca0a6156-1f24-4a5c-867d-f35a33c87876", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.860352] env[63197]: DEBUG oslo_concurrency.lockutils [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] Acquired lock "refresh_cache-72f9b0c4-69b9-49f2-8665-ff77151883af" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.860352] env[63197]: DEBUG nova.network.neutron [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Refreshing network info cache for port ca0a6156-1f24-4a5c-867d-f35a33c87876 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.860352] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:4a:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca0a6156-1f24-4a5c-867d-f35a33c87876', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.869124] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Creating folder: Project (9b568f78ebda4dbaafefa681900293e7). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.872855] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-41bfa8d6-c77c-4736-a091-a0ecfd9773c9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.885020] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Created folder: Project (9b568f78ebda4dbaafefa681900293e7) in parent group-v290286. [ 872.885020] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Creating folder: Instances. Parent ref: group-v290319. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.885020] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4be887ea-7088-40d5-a07d-3946332ac005 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.893777] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Created folder: Instances in parent group-v290319. [ 872.894277] env[63197]: DEBUG oslo.service.loopingcall [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.894623] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.894972] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc756584-3059-4aff-9063-3fd0a79aee46 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.917021] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.917021] env[63197]: value = "task-1364138" [ 872.917021] env[63197]: _type = "Task" [ 872.917021] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.923275] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364138, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.143030] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.143030] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.144027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.759s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.148371] env[63197]: INFO nova.compute.claims [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.239116] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364134, 'name': ReconfigVM_Task, 'duration_secs': 0.279328} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.239550] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 9c1caa2b-c369-425a-8726-cddadf06f338/9c1caa2b-c369-425a-8726-cddadf06f338.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.243392] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5796d12-2f3c-4be6-a0fb-fa8eee5a09c4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.251223] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364135, 'name': Rename_Task, 'duration_secs': 0.139099} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.252696] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.253146] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 873.253146] env[63197]: value = "task-1364139" [ 873.253146] env[63197]: _type = "Task" [ 873.253146] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.253370] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b36eb60-6659-4875-a7dc-c02e276e568b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.263190] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364139, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.264560] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 873.264560] env[63197]: value = "task-1364140" [ 873.264560] env[63197]: _type = "Task" [ 873.264560] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.268210] env[63197]: DEBUG nova.network.neutron [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Updated VIF entry in instance network info cache for port ca0a6156-1f24-4a5c-867d-f35a33c87876. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.268659] env[63197]: DEBUG nova.network.neutron [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Updating instance_info_cache with network_info: [{"id": "ca0a6156-1f24-4a5c-867d-f35a33c87876", "address": "fa:16:3e:a9:4a:d2", "network": {"id": "e437d47b-36c0-481d-acb7-fc82b08ab41e", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1625379393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9b568f78ebda4dbaafefa681900293e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa0d7fd6-4fd3-4451-9208-8a3cf25e30e4", "external-id": "nsx-vlan-transportzone-545", "segmentation_id": 545, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca0a6156-1f", "ovs_interfaceid": "ca0a6156-1f24-4a5c-867d-f35a33c87876", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.275625] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364140, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.300064] env[63197]: DEBUG nova.network.neutron [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Successfully updated port: eae8fdec-5c12-41cd-b92e-8673e5d68cb0 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.426259] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364138, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.655434] env[63197]: DEBUG nova.compute.utils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.659506] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.659600] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 873.707708] env[63197]: DEBUG nova.policy [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccf7a4049c9d4341a59b7a050a1d5f09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d78a0bec6e64f7dacabbf83f0b1b13f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.724385] env[63197]: DEBUG nova.compute.manager [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Received event network-vif-plugged-eae8fdec-5c12-41cd-b92e-8673e5d68cb0 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.724385] env[63197]: DEBUG oslo_concurrency.lockutils [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] Acquiring lock "a0a40337-d8d7-448b-afff-f6849e9d37a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.724681] env[63197]: DEBUG oslo_concurrency.lockutils [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.725088] env[63197]: DEBUG oslo_concurrency.lockutils [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.725283] env[63197]: DEBUG nova.compute.manager [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] No waiting events found dispatching network-vif-plugged-eae8fdec-5c12-41cd-b92e-8673e5d68cb0 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.725610] env[63197]: WARNING nova.compute.manager [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Received unexpected event network-vif-plugged-eae8fdec-5c12-41cd-b92e-8673e5d68cb0 for instance with vm_state building and task_state spawning. [ 873.725933] env[63197]: DEBUG nova.compute.manager [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Received event network-changed-eae8fdec-5c12-41cd-b92e-8673e5d68cb0 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.726238] env[63197]: DEBUG nova.compute.manager [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Refreshing instance network info cache due to event network-changed-eae8fdec-5c12-41cd-b92e-8673e5d68cb0. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.726875] env[63197]: DEBUG oslo_concurrency.lockutils [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] Acquiring lock "refresh_cache-a0a40337-d8d7-448b-afff-f6849e9d37a1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.726875] env[63197]: DEBUG oslo_concurrency.lockutils [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] Acquired lock "refresh_cache-a0a40337-d8d7-448b-afff-f6849e9d37a1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.727052] env[63197]: DEBUG nova.network.neutron [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Refreshing network info cache for port eae8fdec-5c12-41cd-b92e-8673e5d68cb0 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.764552] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364139, 'name': Rename_Task, 'duration_secs': 0.140807} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.764767] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.765308] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99ccc09b-6765-4466-94bd-c55eb424000e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.772010] env[63197]: DEBUG oslo_concurrency.lockutils [req-8ead7a28-e1e9-4c0b-8dfb-ef83cc28dcdf req-78585b2c-0ded-47f7-8f62-49645d7b8c9b service nova] Releasing lock "refresh_cache-72f9b0c4-69b9-49f2-8665-ff77151883af" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.776256] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364140, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.777786] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 873.777786] env[63197]: value = "task-1364141" [ 873.777786] env[63197]: _type = "Task" [ 873.777786] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.786544] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364141, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.803547] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "refresh_cache-a0a40337-d8d7-448b-afff-f6849e9d37a1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.925478] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364138, 'name': CreateVM_Task, 'duration_secs': 0.534023} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.925622] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.926376] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.926463] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.926972] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.927059] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29a30e82-3302-4872-891e-b1c90d25129f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.931864] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 873.931864] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528161c4-c148-b842-8613-b5f6e989d41a" [ 873.931864] env[63197]: _type = "Task" [ 873.931864] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.940563] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528161c4-c148-b842-8613-b5f6e989d41a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.047521] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Successfully created port: c3d0ad10-09c1-4552-b5c0-5a79d8c20724 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.160119] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 874.265771] env[63197]: DEBUG nova.network.neutron [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.278104] env[63197]: DEBUG oslo_vmware.api [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364140, 'name': PowerOnVM_Task, 'duration_secs': 0.573402} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.279338] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.279571] env[63197]: DEBUG nova.compute.manager [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.280968] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f91b738-861f-4473-b425-946170ff99a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.297055] env[63197]: DEBUG oslo_vmware.api [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364141, 'name': PowerOnVM_Task, 'duration_secs': 0.461203} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.299865] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.300840] env[63197]: INFO nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Took 8.05 seconds to spawn the instance on the hypervisor. [ 874.300840] env[63197]: DEBUG nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.301201] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35067df2-f115-45fb-a6ac-226f100d7eb1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.371639] env[63197]: DEBUG nova.network.neutron [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.450753] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528161c4-c148-b842-8613-b5f6e989d41a, 'name': SearchDatastore_Task, 'duration_secs': 0.009838} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.453163] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.453587] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.453930] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.454171] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.454452] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.455765] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946db18b-fb65-43a6-adfe-deb29f723a23 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.461517] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b58f356-bcf6-422d-9349-75fbb0b5ff29 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.468294] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ba6eab-92ea-4f29-bb02-1c29cb1c3f80 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.472284] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.472485] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.473481] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acb3b166-8803-4c36-837c-738a956df6bf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.501412] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741bc033-b06d-4578-a89d-1fde51e1319a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.505985] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 874.505985] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52276c5f-67ad-9748-daad-109de01c02c0" [ 874.505985] env[63197]: _type = "Task" [ 874.505985] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.512963] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef486cd-dad9-44f0-b75f-aed836da1b8c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.520471] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52276c5f-67ad-9748-daad-109de01c02c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.529931] env[63197]: DEBUG nova.compute.provider_tree [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.807757] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.824960] env[63197]: INFO nova.compute.manager [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Took 26.89 seconds to build instance. [ 874.874737] env[63197]: DEBUG oslo_concurrency.lockutils [req-5156f70c-60bd-4a46-8f3d-0aa9c434a01b req-38f157fa-b6c3-4ade-b0e6-8c0bc082769a service nova] Releasing lock "refresh_cache-a0a40337-d8d7-448b-afff-f6849e9d37a1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.875494] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquired lock "refresh_cache-a0a40337-d8d7-448b-afff-f6849e9d37a1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.875655] env[63197]: DEBUG nova.network.neutron [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.020050] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52276c5f-67ad-9748-daad-109de01c02c0, 'name': SearchDatastore_Task, 'duration_secs': 0.015845} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.020703] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-593c3b51-6a02-49b4-b679-621799911177 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.025758] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 875.025758] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528be334-a97b-9496-016d-808584c67d7d" [ 875.025758] env[63197]: _type = "Task" [ 875.025758] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.034300] env[63197]: DEBUG nova.scheduler.client.report [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.037962] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528be334-a97b-9496-016d-808584c67d7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.171408] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.197801] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.198114] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.198302] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.198566] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.198787] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.198980] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.199250] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.199430] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.199609] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.199784] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.199964] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.200839] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db6a45e-9107-4783-8763-e19aca790d6e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.208950] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edce2973-199f-4668-810f-849b25272647 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.327567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c9b79e45-2dc3-47d9-b3ac-0915cd257f54 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "9c1caa2b-c369-425a-8726-cddadf06f338" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.697s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.437317] env[63197]: DEBUG nova.network.neutron [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.536872] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528be334-a97b-9496-016d-808584c67d7d, 'name': SearchDatastore_Task, 'duration_secs': 0.01939} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.537148] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.537409] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 72f9b0c4-69b9-49f2-8665-ff77151883af/72f9b0c4-69b9-49f2-8665-ff77151883af.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.537674] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a943644c-b2c3-433d-9235-de44c396d493 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.542315] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.542801] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.545260] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.779s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.546778] env[63197]: INFO nova.compute.claims [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.555014] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 875.555014] env[63197]: value = "task-1364142" [ 875.555014] env[63197]: _type = "Task" [ 875.555014] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.561984] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364142, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.705417] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Successfully updated port: c3d0ad10-09c1-4552-b5c0-5a79d8c20724 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.793452] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.793731] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.793955] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.794154] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.794323] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.796717] env[63197]: INFO nova.compute.manager [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Terminating instance [ 875.798468] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "refresh_cache-3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.798620] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquired lock "refresh_cache-3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.798784] env[63197]: DEBUG nova.network.neutron [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.830355] env[63197]: DEBUG nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 875.910851] env[63197]: DEBUG nova.compute.manager [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Received event network-vif-plugged-c3d0ad10-09c1-4552-b5c0-5a79d8c20724 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.912180] env[63197]: DEBUG oslo_concurrency.lockutils [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] Acquiring lock "46fd7dec-588e-430e-b51f-9b61d9b148b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.912754] env[63197]: DEBUG oslo_concurrency.lockutils [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.912994] env[63197]: DEBUG oslo_concurrency.lockutils [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.913206] env[63197]: DEBUG nova.compute.manager [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] No waiting events found dispatching network-vif-plugged-c3d0ad10-09c1-4552-b5c0-5a79d8c20724 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.913376] env[63197]: WARNING nova.compute.manager [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Received unexpected event network-vif-plugged-c3d0ad10-09c1-4552-b5c0-5a79d8c20724 for instance with vm_state building and task_state spawning. [ 875.913537] env[63197]: DEBUG nova.compute.manager [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Received event network-changed-c3d0ad10-09c1-4552-b5c0-5a79d8c20724 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.913835] env[63197]: DEBUG nova.compute.manager [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Refreshing instance network info cache due to event network-changed-c3d0ad10-09c1-4552-b5c0-5a79d8c20724. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.913920] env[63197]: DEBUG oslo_concurrency.lockutils [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] Acquiring lock "refresh_cache-46fd7dec-588e-430e-b51f-9b61d9b148b6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.913998] env[63197]: DEBUG oslo_concurrency.lockutils [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] Acquired lock "refresh_cache-46fd7dec-588e-430e-b51f-9b61d9b148b6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.914396] env[63197]: DEBUG nova.network.neutron [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Refreshing network info cache for port c3d0ad10-09c1-4552-b5c0-5a79d8c20724 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.052019] env[63197]: DEBUG nova.compute.utils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.055959] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.056200] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 876.069638] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364142, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.102630] env[63197]: DEBUG nova.policy [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ccf7a4049c9d4341a59b7a050a1d5f09', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d78a0bec6e64f7dacabbf83f0b1b13f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.112607] env[63197]: DEBUG nova.network.neutron [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Updating instance_info_cache with network_info: [{"id": "eae8fdec-5c12-41cd-b92e-8673e5d68cb0", "address": "fa:16:3e:8c:cb:60", "network": {"id": "f014fa77-21b7-4919-a45e-fe413e815fc9", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1113040537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "153c4635bb6740308040698319aff9be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba6157eb-73cb-428a-9f46-99081165d7eb", "external-id": "nsx-vlan-transportzone-463", "segmentation_id": 463, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeae8fdec-5c", "ovs_interfaceid": "eae8fdec-5c12-41cd-b92e-8673e5d68cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.207694] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "refresh_cache-46fd7dec-588e-430e-b51f-9b61d9b148b6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.327957] env[63197]: DEBUG nova.network.neutron [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.359611] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.380909] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Successfully created port: 619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.470607] env[63197]: DEBUG nova.network.neutron [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.476642] env[63197]: DEBUG nova.network.neutron [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.556570] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.573706] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364142, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52869} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.575027] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 72f9b0c4-69b9-49f2-8665-ff77151883af/72f9b0c4-69b9-49f2-8665-ff77151883af.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 876.575027] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.575027] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cdb1e81c-e788-4756-bf20-ca7587bf10b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.581819] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 876.581819] env[63197]: value = "task-1364143" [ 876.581819] env[63197]: _type = "Task" [ 876.581819] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.599957] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364143, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.617828] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Releasing lock "refresh_cache-a0a40337-d8d7-448b-afff-f6849e9d37a1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.617828] env[63197]: DEBUG nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Instance network_info: |[{"id": "eae8fdec-5c12-41cd-b92e-8673e5d68cb0", "address": "fa:16:3e:8c:cb:60", "network": {"id": "f014fa77-21b7-4919-a45e-fe413e815fc9", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1113040537-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "153c4635bb6740308040698319aff9be", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba6157eb-73cb-428a-9f46-99081165d7eb", "external-id": "nsx-vlan-transportzone-463", "segmentation_id": 463, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeae8fdec-5c", "ovs_interfaceid": "eae8fdec-5c12-41cd-b92e-8673e5d68cb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.618185] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:cb:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ba6157eb-73cb-428a-9f46-99081165d7eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eae8fdec-5c12-41cd-b92e-8673e5d68cb0', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.624272] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Creating folder: Project (153c4635bb6740308040698319aff9be). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.627179] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea68b6a7-68dc-4a7e-a344-77056bcc81ec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.642559] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Created folder: Project (153c4635bb6740308040698319aff9be) in parent group-v290286. [ 876.642722] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Creating folder: Instances. Parent ref: group-v290322. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.648275] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b063687-b794-4a7a-95f5-1decc4b263d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.658663] env[63197]: DEBUG nova.compute.manager [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-changed-1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.658663] env[63197]: DEBUG nova.compute.manager [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing instance network info cache due to event network-changed-1d3f4b96-07ba-4875-9441-753c3b9855a9. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 876.658923] env[63197]: DEBUG oslo_concurrency.lockutils [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.659125] env[63197]: DEBUG oslo_concurrency.lockutils [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.659284] env[63197]: DEBUG nova.network.neutron [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing network info cache for port 1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.661474] env[63197]: DEBUG nova.network.neutron [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.665464] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Created folder: Instances in parent group-v290322. [ 876.665464] env[63197]: DEBUG oslo.service.loopingcall [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.668422] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.669095] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55cefba1-d8f6-4c8e-aaa1-fe377748b0ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.695321] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.695321] env[63197]: value = "task-1364146" [ 876.695321] env[63197]: _type = "Task" [ 876.695321] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.704721] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364146, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.885453] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8cf6fc-74ce-4518-ac10-9fe2b88ca3f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.896440] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304b5789-d69d-48d1-8aa3-76626f6a7df1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.933397] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9e34b8-da67-41c4-9b67-811b958b0022 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.943837] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c684581d-a093-46a8-9692-7502037d0518 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.962350] env[63197]: DEBUG nova.compute.provider_tree [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.972900] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Releasing lock "refresh_cache-3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.973359] env[63197]: DEBUG nova.compute.manager [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.973551] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.974687] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c3dd55-47d8-4e4f-8bc3-09af25068772 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.986437] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.986925] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36d2b37a-16eb-4012-bcc7-1507fd8c65b6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.994250] env[63197]: DEBUG oslo_vmware.api [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 876.994250] env[63197]: value = "task-1364147" [ 876.994250] env[63197]: _type = "Task" [ 876.994250] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.007606] env[63197]: DEBUG oslo_vmware.api [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364147, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.091886] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364143, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060969} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.092197] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.093248] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e641b07f-883b-481f-b6ed-f5fc1b348211 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.116094] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 72f9b0c4-69b9-49f2-8665-ff77151883af/72f9b0c4-69b9-49f2-8665-ff77151883af.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.117332] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d3dbb28-2560-4562-b597-d9db65e713a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.136117] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 877.136117] env[63197]: value = "task-1364148" [ 877.136117] env[63197]: _type = "Task" [ 877.136117] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.145380] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.169220] env[63197]: DEBUG oslo_concurrency.lockutils [req-c615ad86-ca1c-4ee1-8de4-453d41dbd347 req-782cc8d8-1d2a-431a-bcf7-0ddd950380ea service nova] Releasing lock "refresh_cache-46fd7dec-588e-430e-b51f-9b61d9b148b6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.171645] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "refresh_cache-46fd7dec-588e-430e-b51f-9b61d9b148b6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.171889] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.210463] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364146, 'name': CreateVM_Task, 'duration_secs': 0.384207} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.210640] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.211372] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.211527] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.211916] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.212188] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-decda1f4-4536-42d9-8fb2-035661c91ffe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.217175] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 877.217175] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52db6aa7-e606-4d8d-409d-39171e4cf7e4" [ 877.217175] env[63197]: _type = "Task" [ 877.217175] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.225204] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52db6aa7-e606-4d8d-409d-39171e4cf7e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.458913] env[63197]: DEBUG nova.network.neutron [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updated VIF entry in instance network info cache for port 1d3f4b96-07ba-4875-9441-753c3b9855a9. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.459307] env[63197]: DEBUG nova.network.neutron [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.463607] env[63197]: DEBUG nova.scheduler.client.report [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.505922] env[63197]: DEBUG oslo_vmware.api [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364147, 'name': PowerOffVM_Task, 'duration_secs': 0.130304} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.505922] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.506075] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.506323] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79144858-f094-465f-a7dd-a8b70715da74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.530198] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.530862] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.530862] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Deleting the datastore file [datastore2] 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.531083] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7119f848-738d-4665-abeb-dd5c8d1514e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.537986] env[63197]: DEBUG oslo_vmware.api [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for the task: (returnval){ [ 877.537986] env[63197]: value = "task-1364150" [ 877.537986] env[63197]: _type = "Task" [ 877.537986] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.546826] env[63197]: DEBUG oslo_vmware.api [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.573273] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 877.599465] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.599798] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.600023] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.600281] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.600553] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.600758] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.601051] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.601260] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.601475] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.601689] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.601905] env[63197]: DEBUG nova.virt.hardware [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.602820] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8b580b-453a-46b3-b41c-88f1228fd0d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.610994] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a1b42b-8cf2-4e74-8310-a5d38c34694b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.645824] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364148, 'name': ReconfigVM_Task, 'duration_secs': 0.307848} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.646153] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 72f9b0c4-69b9-49f2-8665-ff77151883af/72f9b0c4-69b9-49f2-8665-ff77151883af.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.646780] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-142a7425-43ca-4f12-9a8f-31c481afa4b9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.653414] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 877.653414] env[63197]: value = "task-1364151" [ 877.653414] env[63197]: _type = "Task" [ 877.653414] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.661781] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364151, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.730042] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52db6aa7-e606-4d8d-409d-39171e4cf7e4, 'name': SearchDatastore_Task, 'duration_secs': 0.009395} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.730042] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.730042] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.730042] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.730554] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.730554] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.730554] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0f9ae17-a0f1-4b83-b86f-317c43023152 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.741103] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.741540] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.743427] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.747238] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6d5a29c-120c-46ad-ac68-122d35c400d9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.760152] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 877.760152] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5210643a-7eef-966f-b4eb-92d3f9655e4f" [ 877.760152] env[63197]: _type = "Task" [ 877.760152] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.770980] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5210643a-7eef-966f-b4eb-92d3f9655e4f, 'name': SearchDatastore_Task, 'duration_secs': 0.009658} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.772213] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe6e0289-e6bc-4079-b62b-c287708f9c82 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.782061] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 877.782061] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f3af56-50bb-ec38-f958-b1c895835f90" [ 877.782061] env[63197]: _type = "Task" [ 877.782061] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.791890] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f3af56-50bb-ec38-f958-b1c895835f90, 'name': SearchDatastore_Task, 'duration_secs': 0.00863} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.797494] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.797924] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] a0a40337-d8d7-448b-afff-f6849e9d37a1/a0a40337-d8d7-448b-afff-f6849e9d37a1.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.798336] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b813919-196d-4972-8ece-d238d6363627 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.805659] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 877.805659] env[63197]: value = "task-1364152" [ 877.805659] env[63197]: _type = "Task" [ 877.805659] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.815517] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364152, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.964945] env[63197]: DEBUG oslo_concurrency.lockutils [req-e49b2dd3-4692-43e3-8792-60d3a89cdc9b req-790efa45-7621-4fed-9d62-5db8189acac9 service nova] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.968564] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.968748] env[63197]: DEBUG nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.972640] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.049s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.974483] env[63197]: INFO nova.compute.claims [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.027181] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Updating instance_info_cache with network_info: [{"id": "c3d0ad10-09c1-4552-b5c0-5a79d8c20724", "address": "fa:16:3e:c9:e6:0e", "network": {"id": "0d18c832-6ebe-4a21-bf1f-ebacfdfb8d9e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1498788943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d78a0bec6e64f7dacabbf83f0b1b13f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3d0ad10-09", "ovs_interfaceid": "c3d0ad10-09c1-4552-b5c0-5a79d8c20724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.062032] env[63197]: DEBUG oslo_vmware.api [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Task: {'id': task-1364150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088963} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.062032] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.062255] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.062455] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.062742] env[63197]: INFO nova.compute.manager [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Took 1.09 seconds to destroy the instance on the hypervisor. [ 878.062937] env[63197]: DEBUG oslo.service.loopingcall [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.063535] env[63197]: DEBUG nova.compute.manager [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 878.063666] env[63197]: DEBUG nova.network.neutron [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.098656] env[63197]: DEBUG nova.network.neutron [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.168951] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364151, 'name': Rename_Task, 'duration_secs': 0.140334} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.169363] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.171682] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3362f35e-0034-4c99-be6a-b3734b487d1e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.180354] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 878.180354] env[63197]: value = "task-1364153" [ 878.180354] env[63197]: _type = "Task" [ 878.180354] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.195950] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364153, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.229463] env[63197]: DEBUG nova.compute.manager [req-ac18190a-d0dc-4175-9e40-5dfd6307aae2 req-76530bca-9df6-4bb0-b515-7d5feceff031 service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Received event network-vif-plugged-619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.229694] env[63197]: DEBUG oslo_concurrency.lockutils [req-ac18190a-d0dc-4175-9e40-5dfd6307aae2 req-76530bca-9df6-4bb0-b515-7d5feceff031 service nova] Acquiring lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.229914] env[63197]: DEBUG oslo_concurrency.lockutils [req-ac18190a-d0dc-4175-9e40-5dfd6307aae2 req-76530bca-9df6-4bb0-b515-7d5feceff031 service nova] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.230102] env[63197]: DEBUG oslo_concurrency.lockutils [req-ac18190a-d0dc-4175-9e40-5dfd6307aae2 req-76530bca-9df6-4bb0-b515-7d5feceff031 service nova] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.230269] env[63197]: DEBUG nova.compute.manager [req-ac18190a-d0dc-4175-9e40-5dfd6307aae2 req-76530bca-9df6-4bb0-b515-7d5feceff031 service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] No waiting events found dispatching network-vif-plugged-619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 878.230432] env[63197]: WARNING nova.compute.manager [req-ac18190a-d0dc-4175-9e40-5dfd6307aae2 req-76530bca-9df6-4bb0-b515-7d5feceff031 service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Received unexpected event network-vif-plugged-619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 for instance with vm_state building and task_state spawning. [ 878.303868] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Successfully updated port: 619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.325022] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364152, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.478724] env[63197]: DEBUG nova.compute.utils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.484528] env[63197]: DEBUG nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 878.484528] env[63197]: DEBUG nova.network.neutron [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 878.530168] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "refresh_cache-46fd7dec-588e-430e-b51f-9b61d9b148b6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.530552] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Instance network_info: |[{"id": "c3d0ad10-09c1-4552-b5c0-5a79d8c20724", "address": "fa:16:3e:c9:e6:0e", "network": {"id": "0d18c832-6ebe-4a21-bf1f-ebacfdfb8d9e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1498788943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d78a0bec6e64f7dacabbf83f0b1b13f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc3d0ad10-09", "ovs_interfaceid": "c3d0ad10-09c1-4552-b5c0-5a79d8c20724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.530966] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:e6:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '674802e7-b847-4bef-a7a8-f90ac7a3a0a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c3d0ad10-09c1-4552-b5c0-5a79d8c20724', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.538548] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Creating folder: Project (7d78a0bec6e64f7dacabbf83f0b1b13f). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.540920] env[63197]: DEBUG nova.policy [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6380f852eb614c4fb8a82872d6bf97c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '104c325de57847fd9d55ccdb10b8af97', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 878.541706] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9266002e-e20a-4e49-9c30-85ae072f21c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.552143] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Created folder: Project (7d78a0bec6e64f7dacabbf83f0b1b13f) in parent group-v290286. [ 878.553168] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Creating folder: Instances. Parent ref: group-v290325. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.553168] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9132bf28-b988-4092-8f37-8a2a29564e49 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.562444] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Created folder: Instances in parent group-v290325. [ 878.562594] env[63197]: DEBUG oslo.service.loopingcall [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.562785] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.563015] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db098a81-afe5-4d0c-86a9-c638f96aa4ce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.584767] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.584767] env[63197]: value = "task-1364156" [ 878.584767] env[63197]: _type = "Task" [ 878.584767] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.594472] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364156, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.602047] env[63197]: DEBUG nova.network.neutron [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.691443] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364153, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.807860] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "refresh_cache-8500a7e9-3fdc-411e-a48e-189d4d7bffba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.807860] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "refresh_cache-8500a7e9-3fdc-411e-a48e-189d4d7bffba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.807860] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.818784] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364152, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.616912} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.820094] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] a0a40337-d8d7-448b-afff-f6849e9d37a1/a0a40337-d8d7-448b-afff-f6849e9d37a1.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.820592] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.821235] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dccc0c07-0718-440f-a038-ee3a9181e8ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.828362] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 878.828362] env[63197]: value = "task-1364157" [ 878.828362] env[63197]: _type = "Task" [ 878.828362] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.837033] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364157, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.987044] env[63197]: DEBUG nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 879.096880] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364156, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.108305] env[63197]: INFO nova.compute.manager [-] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Took 1.04 seconds to deallocate network for instance. [ 879.128879] env[63197]: DEBUG nova.network.neutron [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Successfully created port: ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 879.191762] env[63197]: DEBUG oslo_vmware.api [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364153, 'name': PowerOnVM_Task, 'duration_secs': 0.53274} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.192463] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.192676] env[63197]: INFO nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Took 8.80 seconds to spawn the instance on the hypervisor. [ 879.192861] env[63197]: DEBUG nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.193682] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d075b49-791e-41df-ae0b-aa8612570787 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.303140] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd03cee-b954-459d-a88f-35b22ab3edfc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.314677] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666cbc0a-031d-462c-83d5-a36590e2550e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.353971] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea04febf-2159-4fbf-a951-3aae9999bb94 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.361811] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364157, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070406} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.364101] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.365149] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ff7a30-2e9b-4448-94bb-4f79a5fae960 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.368082] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c58ba67-1eeb-4151-aa01-9459f511cf07 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.383169] env[63197]: DEBUG nova.compute.provider_tree [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.385951] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 879.407803] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] a0a40337-d8d7-448b-afff-f6849e9d37a1/a0a40337-d8d7-448b-afff-f6849e9d37a1.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.409392] env[63197]: DEBUG nova.scheduler.client.report [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.412609] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78cb731d-a7ce-4ad8-a750-ba63b775660d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.433506] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 879.433506] env[63197]: value = "task-1364158" [ 879.433506] env[63197]: _type = "Task" [ 879.433506] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.444356] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364158, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.595615] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364156, 'name': CreateVM_Task, 'duration_secs': 0.57666} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.595938] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.596626] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.596797] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.597119] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.597592] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-def0a096-ad7c-4515-beca-affb1cd39eb7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.602575] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 879.602575] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bba9da-43d3-d86c-2d9f-29d5a58857ae" [ 879.602575] env[63197]: _type = "Task" [ 879.602575] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.611186] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bba9da-43d3-d86c-2d9f-29d5a58857ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.615231] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.649960] env[63197]: DEBUG nova.network.neutron [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Updating instance_info_cache with network_info: [{"id": "619ff4d4-1f0d-4d8e-87be-1cdcf30874a5", "address": "fa:16:3e:7f:d5:9b", "network": {"id": "0d18c832-6ebe-4a21-bf1f-ebacfdfb8d9e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1498788943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d78a0bec6e64f7dacabbf83f0b1b13f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap619ff4d4-1f", "ovs_interfaceid": "619ff4d4-1f0d-4d8e-87be-1cdcf30874a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.719503] env[63197]: INFO nova.compute.manager [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Took 27.70 seconds to build instance. [ 879.929475] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.956s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.929966] env[63197]: DEBUG nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.932961] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.927s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.934498] env[63197]: INFO nova.compute.claims [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.946739] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364158, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.999286] env[63197]: DEBUG nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 880.023272] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.023511] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.023665] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.023843] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.023988] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.024149] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.024353] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.024507] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.024670] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.024828] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.025010] env[63197]: DEBUG nova.virt.hardware [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.026639] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b171121-e849-4ba7-9976-19538a02ae3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.034973] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ae9b1b-4939-487b-9cf1-7666d193c74e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.114962] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bba9da-43d3-d86c-2d9f-29d5a58857ae, 'name': SearchDatastore_Task, 'duration_secs': 0.022775} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.115330] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.115598] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.115864] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.116174] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.116390] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.116685] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc580822-0769-47d4-a5f0-a54b75902e51 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.126047] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.126268] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.127323] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14483778-48f1-4ac7-be01-2d4d72d890f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.133151] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 880.133151] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5299b010-2256-0341-34ff-bdcf2888bbed" [ 880.133151] env[63197]: _type = "Task" [ 880.133151] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.140449] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5299b010-2256-0341-34ff-bdcf2888bbed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.153963] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "refresh_cache-8500a7e9-3fdc-411e-a48e-189d4d7bffba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.154336] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Instance network_info: |[{"id": "619ff4d4-1f0d-4d8e-87be-1cdcf30874a5", "address": "fa:16:3e:7f:d5:9b", "network": {"id": "0d18c832-6ebe-4a21-bf1f-ebacfdfb8d9e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1498788943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d78a0bec6e64f7dacabbf83f0b1b13f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap619ff4d4-1f", "ovs_interfaceid": "619ff4d4-1f0d-4d8e-87be-1cdcf30874a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.154756] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:d5:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '674802e7-b847-4bef-a7a8-f90ac7a3a0a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '619ff4d4-1f0d-4d8e-87be-1cdcf30874a5', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.163602] env[63197]: DEBUG oslo.service.loopingcall [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.163602] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.163721] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-437298fa-abe1-4953-8b1d-9404fcaac2ea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.187447] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.187447] env[63197]: value = "task-1364159" [ 880.187447] env[63197]: _type = "Task" [ 880.187447] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.195675] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364159, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.222248] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0fd5f065-9fd5-497b-a8ae-caa80f173d38 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.833s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.238310] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.238310] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.265870] env[63197]: DEBUG nova.compute.manager [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Received event network-changed-619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.266081] env[63197]: DEBUG nova.compute.manager [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Refreshing instance network info cache due to event network-changed-619ff4d4-1f0d-4d8e-87be-1cdcf30874a5. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.266311] env[63197]: DEBUG oslo_concurrency.lockutils [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] Acquiring lock "refresh_cache-8500a7e9-3fdc-411e-a48e-189d4d7bffba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.266496] env[63197]: DEBUG oslo_concurrency.lockutils [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] Acquired lock "refresh_cache-8500a7e9-3fdc-411e-a48e-189d4d7bffba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.266608] env[63197]: DEBUG nova.network.neutron [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Refreshing network info cache for port 619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.440136] env[63197]: DEBUG nova.compute.utils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.442822] env[63197]: DEBUG nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.443021] env[63197]: DEBUG nova.network.neutron [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.455699] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364158, 'name': ReconfigVM_Task, 'duration_secs': 0.920172} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.455699] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Reconfigured VM instance instance-00000041 to attach disk [datastore1] a0a40337-d8d7-448b-afff-f6849e9d37a1/a0a40337-d8d7-448b-afff-f6849e9d37a1.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.455699] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af7a217a-72a7-4f59-9935-be5e0b1cfd07 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.463998] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 880.463998] env[63197]: value = "task-1364160" [ 880.463998] env[63197]: _type = "Task" [ 880.463998] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.471757] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364160, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.497998] env[63197]: DEBUG nova.policy [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81fc34f0c73c4baf829c5f9cae7510fc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d992f8d3d334b8783f3dc5eff1cf423', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.643256] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5299b010-2256-0341-34ff-bdcf2888bbed, 'name': SearchDatastore_Task, 'duration_secs': 0.009973} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.644079] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e4d8fcb-0882-4ec3-a8de-29b0e70d3128 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.649242] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 880.649242] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a862a0-0bce-0a57-9afd-b1673100784e" [ 880.649242] env[63197]: _type = "Task" [ 880.649242] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.657582] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a862a0-0bce-0a57-9afd-b1673100784e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.697742] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364159, 'name': CreateVM_Task, 'duration_secs': 0.318443} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.697974] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 880.698655] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.698819] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.699153] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.699401] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f98eac6b-b644-4040-a510-b4f7105f4aa1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.704078] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 880.704078] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5241fb05-43ed-3071-863c-083464aae459" [ 880.704078] env[63197]: _type = "Task" [ 880.704078] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.713610] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5241fb05-43ed-3071-863c-083464aae459, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.725214] env[63197]: DEBUG nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.953862] env[63197]: DEBUG nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.959713] env[63197]: DEBUG nova.network.neutron [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Successfully created port: a42d56b7-bb6b-40d8-a491-2304a1e4a65b {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.978162] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364160, 'name': Rename_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.043522] env[63197]: DEBUG nova.network.neutron [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Updated VIF entry in instance network info cache for port 619ff4d4-1f0d-4d8e-87be-1cdcf30874a5. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.044127] env[63197]: DEBUG nova.network.neutron [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Updating instance_info_cache with network_info: [{"id": "619ff4d4-1f0d-4d8e-87be-1cdcf30874a5", "address": "fa:16:3e:7f:d5:9b", "network": {"id": "0d18c832-6ebe-4a21-bf1f-ebacfdfb8d9e", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1498788943-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d78a0bec6e64f7dacabbf83f0b1b13f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap619ff4d4-1f", "ovs_interfaceid": "619ff4d4-1f0d-4d8e-87be-1cdcf30874a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.059639] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "72f9b0c4-69b9-49f2-8665-ff77151883af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.059885] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.060104] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "72f9b0c4-69b9-49f2-8665-ff77151883af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.060289] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.060458] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.063337] env[63197]: INFO nova.compute.manager [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Terminating instance [ 881.065335] env[63197]: DEBUG nova.compute.manager [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 881.065585] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.066557] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29026767-9914-45e0-b877-bcec89ea195a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.075873] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.076122] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50015c8f-cc59-4d7f-b092-d436c0dbba3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.081919] env[63197]: DEBUG oslo_vmware.api [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 881.081919] env[63197]: value = "task-1364161" [ 881.081919] env[63197]: _type = "Task" [ 881.081919] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.089749] env[63197]: DEBUG oslo_vmware.api [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364161, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.162566] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a862a0-0bce-0a57-9afd-b1673100784e, 'name': SearchDatastore_Task, 'duration_secs': 0.009159} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.164243] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.164551] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 46fd7dec-588e-430e-b51f-9b61d9b148b6/46fd7dec-588e-430e-b51f-9b61d9b148b6.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.164871] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-516778b2-affa-46a9-8775-7398ff160d4d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.172396] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 881.172396] env[63197]: value = "task-1364162" [ 881.172396] env[63197]: _type = "Task" [ 881.172396] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.180576] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364162, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.193689] env[63197]: DEBUG nova.network.neutron [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Successfully updated port: ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 881.218885] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5241fb05-43ed-3071-863c-083464aae459, 'name': SearchDatastore_Task, 'duration_secs': 0.009599} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.219238] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.219498] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.219772] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.219926] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.220139] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.220446] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c47d369-96ba-40ac-b094-43ef6577a548 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.236277] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.237219] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 881.237704] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b90d8c4-6ec5-45dd-9431-6321d5586419 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.245787] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 881.245787] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c01378-8b53-604b-6f5e-d937fd08a871" [ 881.245787] env[63197]: _type = "Task" [ 881.245787] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.250377] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.253882] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c01378-8b53-604b-6f5e-d937fd08a871, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.274671] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee60364-e30a-4641-b52f-e2a02cbba764 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.282960] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d8ad32-03d1-4548-91b3-8ec2d0748e6f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.318494] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c65562-18ba-4db5-b929-1b5955833d05 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.327856] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b83b874-5813-416a-9aed-9b81ac5f70e2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.341826] env[63197]: DEBUG nova.compute.provider_tree [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.476357] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364160, 'name': Rename_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.548184] env[63197]: DEBUG oslo_concurrency.lockutils [req-b44655ad-8c65-494e-85c5-5226051ea66e req-74676e97-5538-4fcc-8459-f095ec96908b service nova] Releasing lock "refresh_cache-8500a7e9-3fdc-411e-a48e-189d4d7bffba" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.598223] env[63197]: DEBUG oslo_vmware.api [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364161, 'name': PowerOffVM_Task, 'duration_secs': 0.246718} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.598733] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.599240] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.599533] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4570feaa-f3a6-4e27-8516-1cf498600ed8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.670455] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.670758] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.670960] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Deleting the datastore file [datastore2] 72f9b0c4-69b9-49f2-8665-ff77151883af {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.671274] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f408fb5-4b73-4255-816c-ba4459c014e1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.681261] env[63197]: DEBUG oslo_vmware.api [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for the task: (returnval){ [ 881.681261] env[63197]: value = "task-1364164" [ 881.681261] env[63197]: _type = "Task" [ 881.681261] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.684158] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364162, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.691034] env[63197]: DEBUG oslo_vmware.api [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.697632] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.697767] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquired lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.697912] env[63197]: DEBUG nova.network.neutron [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.756529] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c01378-8b53-604b-6f5e-d937fd08a871, 'name': SearchDatastore_Task, 'duration_secs': 0.011746} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.757364] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-620feef3-f7fa-4dba-9988-855ad8881926 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.763366] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 881.763366] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525dfc0c-ae18-9b7f-3f1a-3c4e480c5c00" [ 881.763366] env[63197]: _type = "Task" [ 881.763366] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.771268] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525dfc0c-ae18-9b7f-3f1a-3c4e480c5c00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.845776] env[63197]: DEBUG nova.scheduler.client.report [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.972387] env[63197]: DEBUG nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.979790] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364160, 'name': Rename_Task, 'duration_secs': 1.099615} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.980069] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.980313] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e756fa5-80b7-4b3c-9235-365e5494925d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.986672] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 881.986672] env[63197]: value = "task-1364165" [ 881.986672] env[63197]: _type = "Task" [ 881.986672] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.998050] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364165, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.000713] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.000713] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.000883] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.001102] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.001194] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.001340] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.001544] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.001699] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.001908] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.002084] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.002256] env[63197]: DEBUG nova.virt.hardware [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.003079] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bfcc17-d60c-4762-8dec-f048846919ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.010179] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a769c1df-e9c7-45f9-95a5-dc5bad0d0345 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.183738] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364162, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521861} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.184105] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 46fd7dec-588e-430e-b51f-9b61d9b148b6/46fd7dec-588e-430e-b51f-9b61d9b148b6.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.184377] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.184530] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c68ba051-c8d4-43f2-94e1-25fe41aca18e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.195073] env[63197]: DEBUG oslo_vmware.api [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Task: {'id': task-1364164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148079} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.196165] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.196359] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.196535] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.196703] env[63197]: INFO nova.compute.manager [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Took 1.13 seconds to destroy the instance on the hypervisor. [ 882.196929] env[63197]: DEBUG oslo.service.loopingcall [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.197197] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 882.197197] env[63197]: value = "task-1364166" [ 882.197197] env[63197]: _type = "Task" [ 882.197197] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.197387] env[63197]: DEBUG nova.compute.manager [-] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.197520] env[63197]: DEBUG nova.network.neutron [-] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.233728] env[63197]: DEBUG nova.network.neutron [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 882.275803] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525dfc0c-ae18-9b7f-3f1a-3c4e480c5c00, 'name': SearchDatastore_Task, 'duration_secs': 0.009304} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.278432] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.279043] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 8500a7e9-3fdc-411e-a48e-189d4d7bffba/8500a7e9-3fdc-411e-a48e-189d4d7bffba.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 882.279340] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f000551e-88f6-40e6-b373-a1d4e5c54020 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.285797] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 882.285797] env[63197]: value = "task-1364167" [ 882.285797] env[63197]: _type = "Task" [ 882.285797] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.293783] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364167, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.295934] env[63197]: DEBUG nova.compute.manager [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Received event network-vif-plugged-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.296060] env[63197]: DEBUG oslo_concurrency.lockutils [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.296227] env[63197]: DEBUG oslo_concurrency.lockutils [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.296394] env[63197]: DEBUG oslo_concurrency.lockutils [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.296564] env[63197]: DEBUG nova.compute.manager [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] No waiting events found dispatching network-vif-plugged-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 882.296747] env[63197]: WARNING nova.compute.manager [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Received unexpected event network-vif-plugged-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 for instance with vm_state building and task_state spawning. [ 882.296915] env[63197]: DEBUG nova.compute.manager [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Received event network-changed-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.297107] env[63197]: DEBUG nova.compute.manager [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Refreshing instance network info cache due to event network-changed-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 882.297325] env[63197]: DEBUG oslo_concurrency.lockutils [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] Acquiring lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.350931] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.352175] env[63197]: DEBUG nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.355425] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.521s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.360770] env[63197]: INFO nova.compute.claims [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.438768] env[63197]: DEBUG nova.network.neutron [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updating instance_info_cache with network_info: [{"id": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "address": "fa:16:3e:c3:5f:0c", "network": {"id": "138b731d-9b6e-45ea-b051-6a7e5c8a2471", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1660823209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104c325de57847fd9d55ccdb10b8af97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d2ab9-dd", "ovs_interfaceid": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.499314] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364165, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.712804] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089424} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.712804] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.713574] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c75ffee-03b3-42bc-b6e8-51a85d590746 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.739557] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 46fd7dec-588e-430e-b51f-9b61d9b148b6/46fd7dec-588e-430e-b51f-9b61d9b148b6.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.740406] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8402d948-be88-4336-9fbf-504a3ce95795 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.761407] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 882.761407] env[63197]: value = "task-1364168" [ 882.761407] env[63197]: _type = "Task" [ 882.761407] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.769767] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364168, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.795622] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364167, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478679} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.795880] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 8500a7e9-3fdc-411e-a48e-189d4d7bffba/8500a7e9-3fdc-411e-a48e-189d4d7bffba.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.796108] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.796378] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0111eaab-a819-4694-a1dc-9b9aa3e5fc18 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.803150] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 882.803150] env[63197]: value = "task-1364169" [ 882.803150] env[63197]: _type = "Task" [ 882.803150] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.811339] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364169, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.821624] env[63197]: DEBUG nova.network.neutron [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Successfully updated port: a42d56b7-bb6b-40d8-a491-2304a1e4a65b {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.862373] env[63197]: DEBUG nova.compute.utils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.867024] env[63197]: DEBUG nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 882.867024] env[63197]: DEBUG nova.network.neutron [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 882.915595] env[63197]: DEBUG nova.policy [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b50be4b75a94b4481c9c65ea1e4e9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bb89fb32d8c4726a9a3104d68ce560a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 882.942881] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Releasing lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.942881] env[63197]: DEBUG nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Instance network_info: |[{"id": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "address": "fa:16:3e:c3:5f:0c", "network": {"id": "138b731d-9b6e-45ea-b051-6a7e5c8a2471", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1660823209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104c325de57847fd9d55ccdb10b8af97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d2ab9-dd", "ovs_interfaceid": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 882.943099] env[63197]: DEBUG oslo_concurrency.lockutils [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] Acquired lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.943099] env[63197]: DEBUG nova.network.neutron [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Refreshing network info cache for port ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.943980] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:5f:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69f65356-c85e-4b7f-ad28-7c7b5e8cf50c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.952259] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Creating folder: Project (104c325de57847fd9d55ccdb10b8af97). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.955553] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fbeebcfe-7477-47b7-908d-aa0784adef93 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.967276] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Created folder: Project (104c325de57847fd9d55ccdb10b8af97) in parent group-v290286. [ 882.967416] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Creating folder: Instances. Parent ref: group-v290329. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.967646] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff683632-ab5f-4364-8a98-3bff05421e1b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.976604] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Created folder: Instances in parent group-v290329. [ 882.976839] env[63197]: DEBUG oslo.service.loopingcall [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.977036] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.977243] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29e2b72a-b9c6-4b47-a0e5-0c7774ba875f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.003313] env[63197]: DEBUG oslo_vmware.api [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364165, 'name': PowerOnVM_Task, 'duration_secs': 0.845072} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.004186] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.004395] env[63197]: INFO nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Took 10.25 seconds to spawn the instance on the hypervisor. [ 883.004571] env[63197]: DEBUG nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.004819] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.004819] env[63197]: value = "task-1364172" [ 883.004819] env[63197]: _type = "Task" [ 883.004819] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.007367] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6a71bc-b85a-4c77-b2c8-cf5d141e44dd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.025632] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364172, 'name': CreateVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.158833] env[63197]: DEBUG nova.network.neutron [-] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.271930] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364168, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.297724] env[63197]: DEBUG nova.network.neutron [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updated VIF entry in instance network info cache for port ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 883.298155] env[63197]: DEBUG nova.network.neutron [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updating instance_info_cache with network_info: [{"id": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "address": "fa:16:3e:c3:5f:0c", "network": {"id": "138b731d-9b6e-45ea-b051-6a7e5c8a2471", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1660823209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104c325de57847fd9d55ccdb10b8af97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d2ab9-dd", "ovs_interfaceid": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.315557] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364169, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070128} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.315557] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 883.315557] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f120c73-6cd5-442e-92ee-0efe144fcd73 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.331833] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.331833] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.331833] env[63197]: DEBUG nova.network.neutron [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.342029] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 8500a7e9-3fdc-411e-a48e-189d4d7bffba/8500a7e9-3fdc-411e-a48e-189d4d7bffba.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 883.342620] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a8241ef-c458-498c-aee1-a10841baa50b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.358976] env[63197]: DEBUG nova.network.neutron [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Successfully created port: 088f6f2e-acc0-4966-9ee5-47b6c6799c5b {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 883.366505] env[63197]: DEBUG nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.369375] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 883.369375] env[63197]: value = "task-1364173" [ 883.369375] env[63197]: _type = "Task" [ 883.369375] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.382037] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364173, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.520907] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364172, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.534799] env[63197]: INFO nova.compute.manager [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Took 27.34 seconds to build instance. [ 883.662488] env[63197]: INFO nova.compute.manager [-] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Took 1.46 seconds to deallocate network for instance. [ 883.664816] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5dc45bb-cbdd-44bb-901b-9f8815f7c491 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.676399] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b61f75-70d2-4f40-af18-25c16466bca9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.708109] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3652e0-9477-415e-95e0-03500d3a2090 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.717278] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d457b405-7169-4b17-96a0-6c6bde3d6790 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.734160] env[63197]: DEBUG nova.compute.provider_tree [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.776246] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364168, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.800871] env[63197]: DEBUG oslo_concurrency.lockutils [req-bf0f9947-018d-4589-bb51-abd79e228057 req-7eaf1400-c68e-4f2d-b2a1-11e0e19d91fd service nova] Releasing lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.883808] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364173, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.912520] env[63197]: DEBUG nova.network.neutron [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.022354] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364172, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.037767] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4d163d32-2436-496d-81fa-58d8fe7b3055 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.577s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.074352] env[63197]: DEBUG nova.network.neutron [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Updating instance_info_cache with network_info: [{"id": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "address": "fa:16:3e:18:1d:83", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa42d56b7-bb", "ovs_interfaceid": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.172369] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.238437] env[63197]: DEBUG nova.scheduler.client.report [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.259443] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "a0a40337-d8d7-448b-afff-f6849e9d37a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.259801] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.260118] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "a0a40337-d8d7-448b-afff-f6849e9d37a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.260383] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.260637] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.263614] env[63197]: INFO nova.compute.manager [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Terminating instance [ 884.265883] env[63197]: DEBUG nova.compute.manager [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 884.266173] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 884.267389] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09aa6e18-2425-4c04-af91-aae3ec6727b2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.282447] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364168, 'name': ReconfigVM_Task, 'duration_secs': 1.101116} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.285337] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 46fd7dec-588e-430e-b51f-9b61d9b148b6/46fd7dec-588e-430e-b51f-9b61d9b148b6.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.286171] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 884.286467] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16d252a5-7e3e-4d9e-b23e-0fa2192a94ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.288383] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c61ed0f-1377-4c59-ba81-f644306f77bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.295044] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 884.295044] env[63197]: value = "task-1364174" [ 884.295044] env[63197]: _type = "Task" [ 884.295044] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.296524] env[63197]: DEBUG oslo_vmware.api [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 884.296524] env[63197]: value = "task-1364175" [ 884.296524] env[63197]: _type = "Task" [ 884.296524] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.309634] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364174, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.312834] env[63197]: DEBUG oslo_vmware.api [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.327390] env[63197]: DEBUG nova.compute.manager [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Received event network-vif-deleted-ca0a6156-1f24-4a5c-867d-f35a33c87876 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.327591] env[63197]: DEBUG nova.compute.manager [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Received event network-vif-plugged-a42d56b7-bb6b-40d8-a491-2304a1e4a65b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.327779] env[63197]: DEBUG oslo_concurrency.lockutils [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] Acquiring lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.328045] env[63197]: DEBUG oslo_concurrency.lockutils [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.328273] env[63197]: DEBUG oslo_concurrency.lockutils [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.328484] env[63197]: DEBUG nova.compute.manager [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] No waiting events found dispatching network-vif-plugged-a42d56b7-bb6b-40d8-a491-2304a1e4a65b {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 884.328699] env[63197]: WARNING nova.compute.manager [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Received unexpected event network-vif-plugged-a42d56b7-bb6b-40d8-a491-2304a1e4a65b for instance with vm_state building and task_state spawning. [ 884.328909] env[63197]: DEBUG nova.compute.manager [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Received event network-changed-a42d56b7-bb6b-40d8-a491-2304a1e4a65b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.329123] env[63197]: DEBUG nova.compute.manager [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Refreshing instance network info cache due to event network-changed-a42d56b7-bb6b-40d8-a491-2304a1e4a65b. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.329339] env[63197]: DEBUG oslo_concurrency.lockutils [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] Acquiring lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.380627] env[63197]: DEBUG nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.386614] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364173, 'name': ReconfigVM_Task, 'duration_secs': 0.686843} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.386934] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 8500a7e9-3fdc-411e-a48e-189d4d7bffba/8500a7e9-3fdc-411e-a48e-189d4d7bffba.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.387577] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6608518-9a11-4cf3-919a-9327b4dae56c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.395368] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 884.395368] env[63197]: value = "task-1364176" [ 884.395368] env[63197]: _type = "Task" [ 884.395368] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.407194] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364176, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.415675] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.415937] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.416069] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.416254] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.416392] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.416537] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.416738] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.416897] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.417090] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.417252] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.417420] env[63197]: DEBUG nova.virt.hardware [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.418324] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d98a5b-7de8-439b-86ac-d967c881aaa4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.426172] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4111ed61-faeb-49ad-a5dc-c8d6b33063a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.520626] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364172, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.540301] env[63197]: DEBUG nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 884.577040] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.577270] env[63197]: DEBUG nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Instance network_info: |[{"id": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "address": "fa:16:3e:18:1d:83", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa42d56b7-bb", "ovs_interfaceid": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.577834] env[63197]: DEBUG oslo_concurrency.lockutils [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] Acquired lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.578071] env[63197]: DEBUG nova.network.neutron [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Refreshing network info cache for port a42d56b7-bb6b-40d8-a491-2304a1e4a65b {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.579216] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:1d:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a42d56b7-bb6b-40d8-a491-2304a1e4a65b', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.586479] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Creating folder: Project (0d992f8d3d334b8783f3dc5eff1cf423). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.589095] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aabee3aa-ec77-407c-9751-d635786d08c8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.600284] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Created folder: Project (0d992f8d3d334b8783f3dc5eff1cf423) in parent group-v290286. [ 884.600499] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Creating folder: Instances. Parent ref: group-v290332. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 884.600750] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-17ce43c1-5da9-4f0d-8f1c-e5036977915b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.609655] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Created folder: Instances in parent group-v290332. [ 884.609930] env[63197]: DEBUG oslo.service.loopingcall [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.610156] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.610381] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d73b7ba-8af3-4dd6-8159-4d4a5db92b60 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.628948] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.628948] env[63197]: value = "task-1364179" [ 884.628948] env[63197]: _type = "Task" [ 884.628948] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.638015] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364179, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.745610] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.746216] env[63197]: DEBUG nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.752191] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.487s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.752477] env[63197]: DEBUG nova.objects.instance [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lazy-loading 'resources' on Instance uuid c55b5b2b-5ecc-43bb-a279-7370cd9ac722 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.811442] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364174, 'name': Rename_Task, 'duration_secs': 0.170514} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.814548] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.814836] env[63197]: DEBUG oslo_vmware.api [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364175, 'name': PowerOffVM_Task, 'duration_secs': 0.184766} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.815064] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ad95744-3ed7-4d39-8379-aa845511a2bf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.816825] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.816942] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.817149] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-55d7ead3-f0da-4a55-91e2-50840e7831fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.823850] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 884.823850] env[63197]: value = "task-1364180" [ 884.823850] env[63197]: _type = "Task" [ 884.823850] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.833876] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.882182] env[63197]: DEBUG nova.network.neutron [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Updated VIF entry in instance network info cache for port a42d56b7-bb6b-40d8-a491-2304a1e4a65b. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.882182] env[63197]: DEBUG nova.network.neutron [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Updating instance_info_cache with network_info: [{"id": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "address": "fa:16:3e:18:1d:83", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa42d56b7-bb", "ovs_interfaceid": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.894558] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.894823] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.895028] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Deleting the datastore file [datastore1] a0a40337-d8d7-448b-afff-f6849e9d37a1 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.895352] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ee0f6e4-b6fd-4f41-80ad-8477a5747617 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.906641] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364176, 'name': Rename_Task, 'duration_secs': 0.147163} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.907970] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.908453] env[63197]: DEBUG oslo_vmware.api [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for the task: (returnval){ [ 884.908453] env[63197]: value = "task-1364182" [ 884.908453] env[63197]: _type = "Task" [ 884.908453] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.908677] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37a92128-8ffe-4168-9f2e-a4e781792a9b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.918824] env[63197]: DEBUG oslo_vmware.api [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364182, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.920233] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 884.920233] env[63197]: value = "task-1364183" [ 884.920233] env[63197]: _type = "Task" [ 884.920233] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.927446] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.023028] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364172, 'name': CreateVM_Task, 'duration_secs': 1.851857} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.023215] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.023936] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.024163] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.024498] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.024965] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-637ca062-a248-4577-b71e-0d69066e5315 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.031343] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 885.031343] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52852979-b8e1-57d5-d3fb-a1aea260f11b" [ 885.031343] env[63197]: _type = "Task" [ 885.031343] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.038525] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52852979-b8e1-57d5-d3fb-a1aea260f11b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.073813] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.126518] env[63197]: DEBUG nova.network.neutron [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Successfully updated port: 088f6f2e-acc0-4966-9ee5-47b6c6799c5b {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.138822] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364179, 'name': CreateVM_Task, 'duration_secs': 0.301827} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.139596] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.140266] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.256164] env[63197]: DEBUG nova.compute.utils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.260298] env[63197]: DEBUG nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.260401] env[63197]: DEBUG nova.network.neutron [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.314906] env[63197]: DEBUG nova.policy [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '514b7795ecdc49069e7f3cab700a0ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '527721557235413e99c4a2eaa086486c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.335335] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364180, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.384845] env[63197]: DEBUG oslo_concurrency.lockutils [req-8a334e64-14f7-43e7-9a04-bcd4e6b610f8 req-0116dd72-c6be-415e-94da-ba392d89a459 service nova] Releasing lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.423051] env[63197]: DEBUG oslo_vmware.api [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Task: {'id': task-1364182, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159144} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.426356] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.426560] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.426745] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.426951] env[63197]: INFO nova.compute.manager [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Took 1.16 seconds to destroy the instance on the hypervisor. [ 885.427188] env[63197]: DEBUG oslo.service.loopingcall [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.429605] env[63197]: DEBUG nova.compute.manager [-] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.429705] env[63197]: DEBUG nova.network.neutron [-] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 885.438021] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364183, 'name': PowerOnVM_Task, 'duration_secs': 0.47939} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.438389] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.438709] env[63197]: INFO nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Took 7.86 seconds to spawn the instance on the hypervisor. [ 885.439026] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.440209] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0f9b29-fa26-4b45-8917-2df514f0a457 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.498778] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21691b32-46bf-48e7-a65c-2a58bd602f1a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.507513] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d22f04-10d3-442a-894a-9101f8f6f403 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.546348] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af5f860-f369-47ea-af7c-5f31bf66ba4b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.561620] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-982ad173-eef3-406f-8dfc-92c43b356e44 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.567284] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52852979-b8e1-57d5-d3fb-a1aea260f11b, 'name': SearchDatastore_Task, 'duration_secs': 0.009264} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.567687] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.568018] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.569433] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.569433] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.569433] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.569653] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.570040] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.570397] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8cf68d06-c354-4e68-b493-d5009da0d084 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.584640] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e0492b0-11f0-4885-8e40-01b4f6db65d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.586908] env[63197]: DEBUG nova.compute.provider_tree [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.593013] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 885.593013] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ae4f9c-df33-ebf5-f72e-89cd92af2852" [ 885.593013] env[63197]: _type = "Task" [ 885.593013] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.600272] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.600272] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.601369] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ee27db3-14b5-47ad-8c7f-7b0d0e26c4fa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.609399] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ae4f9c-df33-ebf5-f72e-89cd92af2852, 'name': SearchDatastore_Task, 'duration_secs': 0.008979} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.610170] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.610491] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.610830] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.612591] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 885.612591] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5291e2f8-469f-9742-cc5c-bd97344c4b75" [ 885.612591] env[63197]: _type = "Task" [ 885.612591] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.623026] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5291e2f8-469f-9742-cc5c-bd97344c4b75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.629143] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.629388] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.629652] env[63197]: DEBUG nova.network.neutron [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 885.737132] env[63197]: DEBUG nova.network.neutron [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Successfully created port: e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.761239] env[63197]: DEBUG nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.841630] env[63197]: DEBUG oslo_vmware.api [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364180, 'name': PowerOnVM_Task, 'duration_secs': 0.519089} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.842516] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.842814] env[63197]: INFO nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Took 10.67 seconds to spawn the instance on the hypervisor. [ 885.842997] env[63197]: DEBUG nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.843896] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ffa7d6-88c6-425b-9769-c2785cb044c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.961576] env[63197]: INFO nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Took 23.60 seconds to build instance. [ 886.090946] env[63197]: DEBUG nova.scheduler.client.report [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.127494] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5291e2f8-469f-9742-cc5c-bd97344c4b75, 'name': SearchDatastore_Task, 'duration_secs': 0.00834} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.128368] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efc14382-1bc8-48a6-acec-8995eaf52628 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.135057] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 886.135057] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52954e4f-ec6a-8b5a-ab52-7351c06f58ab" [ 886.135057] env[63197]: _type = "Task" [ 886.135057] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.144645] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52954e4f-ec6a-8b5a-ab52-7351c06f58ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.183146] env[63197]: DEBUG nova.network.neutron [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 886.255015] env[63197]: DEBUG nova.network.neutron [-] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.369643] env[63197]: INFO nova.compute.manager [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Took 24.87 seconds to build instance. [ 886.381306] env[63197]: DEBUG nova.compute.manager [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Received event network-vif-plugged-088f6f2e-acc0-4966-9ee5-47b6c6799c5b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.381524] env[63197]: DEBUG oslo_concurrency.lockutils [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] Acquiring lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.381731] env[63197]: DEBUG oslo_concurrency.lockutils [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.381966] env[63197]: DEBUG oslo_concurrency.lockutils [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.382158] env[63197]: DEBUG nova.compute.manager [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] No waiting events found dispatching network-vif-plugged-088f6f2e-acc0-4966-9ee5-47b6c6799c5b {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.382324] env[63197]: WARNING nova.compute.manager [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Received unexpected event network-vif-plugged-088f6f2e-acc0-4966-9ee5-47b6c6799c5b for instance with vm_state building and task_state spawning. [ 886.382482] env[63197]: DEBUG nova.compute.manager [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Received event network-changed-088f6f2e-acc0-4966-9ee5-47b6c6799c5b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.382632] env[63197]: DEBUG nova.compute.manager [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Refreshing instance network info cache due to event network-changed-088f6f2e-acc0-4966-9ee5-47b6c6799c5b. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.382857] env[63197]: DEBUG oslo_concurrency.lockutils [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] Acquiring lock "refresh_cache-1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.407487] env[63197]: DEBUG nova.network.neutron [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Updating instance_info_cache with network_info: [{"id": "088f6f2e-acc0-4966-9ee5-47b6c6799c5b", "address": "fa:16:3e:80:eb:96", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap088f6f2e-ac", "ovs_interfaceid": "088f6f2e-acc0-4966-9ee5-47b6c6799c5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.465243] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.472s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.595874] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.598099] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.241s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.599754] env[63197]: INFO nova.compute.claims [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 886.622028] env[63197]: INFO nova.scheduler.client.report [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Deleted allocations for instance c55b5b2b-5ecc-43bb-a279-7370cd9ac722 [ 886.647850] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52954e4f-ec6a-8b5a-ab52-7351c06f58ab, 'name': SearchDatastore_Task, 'duration_secs': 0.009178} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.648871] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.649686] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 5b9c6a1c-4212-4941-aa6c-364fb82f5e64/5b9c6a1c-4212-4941-aa6c-364fb82f5e64.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.649686] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.649903] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.650037] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53c82b7d-9ebf-4897-9715-19c913d17647 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.652491] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c50fbbc-6e4d-4d31-bf64-829bed7b0dab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.661932] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.662095] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.663776] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39883ce0-df8c-442d-a8e8-71cc2ad68fcc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.666968] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 886.666968] env[63197]: value = "task-1364184" [ 886.666968] env[63197]: _type = "Task" [ 886.666968] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.671174] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 886.671174] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5258a14a-3738-6eea-5c00-e4e912560754" [ 886.671174] env[63197]: _type = "Task" [ 886.671174] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.679017] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.682960] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5258a14a-3738-6eea-5c00-e4e912560754, 'name': SearchDatastore_Task, 'duration_secs': 0.009698} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.683503] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50cb7299-2ee4-4f00-a434-dbe645dcfded {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.688120] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 886.688120] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ec3e3c-389a-330e-9055-69ac020fdf05" [ 886.688120] env[63197]: _type = "Task" [ 886.688120] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.696046] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ec3e3c-389a-330e-9055-69ac020fdf05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.758155] env[63197]: INFO nova.compute.manager [-] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Took 1.33 seconds to deallocate network for instance. [ 886.775188] env[63197]: DEBUG nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.806804] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.806804] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.806993] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.807075] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.807195] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.807338] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.807566] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.807736] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.811022] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.811022] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.811022] env[63197]: DEBUG nova.virt.hardware [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.811022] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9258f548-d301-4960-82f3-4e001c4dcf9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.818624] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf4a700-38a1-4579-9fc0-4142a5f97dd9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.873284] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0b026968-ee34-4e97-a51e-5b2af1e28475 tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.906s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.910319] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.910550] env[63197]: DEBUG nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Instance network_info: |[{"id": "088f6f2e-acc0-4966-9ee5-47b6c6799c5b", "address": "fa:16:3e:80:eb:96", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap088f6f2e-ac", "ovs_interfaceid": "088f6f2e-acc0-4966-9ee5-47b6c6799c5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 886.910889] env[63197]: DEBUG oslo_concurrency.lockutils [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] Acquired lock "refresh_cache-1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.911091] env[63197]: DEBUG nova.network.neutron [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Refreshing network info cache for port 088f6f2e-acc0-4966-9ee5-47b6c6799c5b {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.912519] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:eb:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '088f6f2e-acc0-4966-9ee5-47b6c6799c5b', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.920653] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Creating folder: Project (8bb89fb32d8c4726a9a3104d68ce560a). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.924283] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4491daf6-070e-4be9-a1f7-51738cbda9e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.938330] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Created folder: Project (8bb89fb32d8c4726a9a3104d68ce560a) in parent group-v290286. [ 886.939465] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Creating folder: Instances. Parent ref: group-v290335. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.939465] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-059715b0-7dbe-48b2-9d3a-b58cfacac3fa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.949734] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Created folder: Instances in parent group-v290335. [ 886.950055] env[63197]: DEBUG oslo.service.loopingcall [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.950328] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 886.950563] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1ffa769-bce3-41fa-8e32-429f78932153 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.971270] env[63197]: DEBUG nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 886.981198] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.981198] env[63197]: value = "task-1364187" [ 886.981198] env[63197]: _type = "Task" [ 886.981198] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.990455] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364187, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.026613] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "46fd7dec-588e-430e-b51f-9b61d9b148b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.026613] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.026613] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "46fd7dec-588e-430e-b51f-9b61d9b148b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.026613] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.026885] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.028890] env[63197]: INFO nova.compute.manager [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Terminating instance [ 887.031154] env[63197]: DEBUG nova.compute.manager [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.031413] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.032410] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897549a7-7a27-4be6-bffd-bd0b65eb4eca {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.040730] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.041246] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ecd0064-c135-4354-84cf-f28a06566bc4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.048661] env[63197]: DEBUG oslo_vmware.api [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 887.048661] env[63197]: value = "task-1364188" [ 887.048661] env[63197]: _type = "Task" [ 887.048661] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.058651] env[63197]: DEBUG oslo_vmware.api [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.089678] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.089678] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.089678] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.089678] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.089941] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.095033] env[63197]: INFO nova.compute.manager [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Terminating instance [ 887.097449] env[63197]: DEBUG nova.compute.manager [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.097707] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.098648] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd57ef7-3041-41c7-91de-c2fc2314cd3e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.113147] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.113147] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-507fb065-d325-48b7-9372-91d986da18ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.119993] env[63197]: DEBUG oslo_vmware.api [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 887.119993] env[63197]: value = "task-1364189" [ 887.119993] env[63197]: _type = "Task" [ 887.119993] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.134241] env[63197]: DEBUG oslo_vmware.api [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364189, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.134876] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4dd517b-f6d9-43fe-bc52-c1c8b5775e91 tempest-ServersNegativeTestMultiTenantJSON-1642908556 tempest-ServersNegativeTestMultiTenantJSON-1642908556-project-member] Lock "c55b5b2b-5ecc-43bb-a279-7370cd9ac722" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.762s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.178944] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484182} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.179411] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 5b9c6a1c-4212-4941-aa6c-364fb82f5e64/5b9c6a1c-4212-4941-aa6c-364fb82f5e64.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.179812] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.180216] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6c40f2e-3175-4959-b1e8-94c2cf00a26c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.189573] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 887.189573] env[63197]: value = "task-1364190" [ 887.189573] env[63197]: _type = "Task" [ 887.189573] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.204736] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ec3e3c-389a-330e-9055-69ac020fdf05, 'name': SearchDatastore_Task, 'duration_secs': 0.007733} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.208067] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.208393] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/9c80bc51-c8f6-436d-9d19-47f6d1aadbf6.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 887.208738] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364190, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.209011] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-29781a12-1b01-4f14-b48c-884a1be21391 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.215329] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 887.215329] env[63197]: value = "task-1364191" [ 887.215329] env[63197]: _type = "Task" [ 887.215329] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.224478] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364191, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.268730] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.333302] env[63197]: DEBUG nova.network.neutron [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Updated VIF entry in instance network info cache for port 088f6f2e-acc0-4966-9ee5-47b6c6799c5b. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.333964] env[63197]: DEBUG nova.network.neutron [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Updating instance_info_cache with network_info: [{"id": "088f6f2e-acc0-4966-9ee5-47b6c6799c5b", "address": "fa:16:3e:80:eb:96", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap088f6f2e-ac", "ovs_interfaceid": "088f6f2e-acc0-4966-9ee5-47b6c6799c5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.499922] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364187, 'name': CreateVM_Task, 'duration_secs': 0.413222} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.499922] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.501060] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.501764] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.502117] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.504713] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.505177] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90f700fb-8e5b-4ae7-9804-583ed4576335 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.513457] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 887.513457] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52059591-0ab7-0646-88b6-3acf1b7ae206" [ 887.513457] env[63197]: _type = "Task" [ 887.513457] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.530016] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52059591-0ab7-0646-88b6-3acf1b7ae206, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.563285] env[63197]: DEBUG oslo_vmware.api [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364188, 'name': PowerOffVM_Task, 'duration_secs': 0.191086} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.563811] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.564120] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.564575] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2051485-b509-4519-ba5d-5a95537d599b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.647503] env[63197]: DEBUG oslo_vmware.api [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364189, 'name': PowerOffVM_Task, 'duration_secs': 0.166679} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.647503] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.647503] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.647503] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.647503] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.647792] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleting the datastore file [datastore1] 46fd7dec-588e-430e-b51f-9b61d9b148b6 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.647792] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c41a6f55-6971-4570-8d76-3d00824b475f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.649301] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60ae13ca-dfd2-4b05-8c8e-473845876dad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.662183] env[63197]: DEBUG oslo_vmware.api [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 887.662183] env[63197]: value = "task-1364194" [ 887.662183] env[63197]: _type = "Task" [ 887.662183] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.674568] env[63197]: DEBUG nova.network.neutron [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Successfully updated port: e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.708712] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364190, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075093} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.708991] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.710066] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaefd17f-c158-431c-bd42-d76ca242b8ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.739284] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 5b9c6a1c-4212-4941-aa6c-364fb82f5e64/5b9c6a1c-4212-4941-aa6c-364fb82f5e64.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.744948] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8288a542-19fa-47fa-931c-174d927d73ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.759791] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.760018] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.760209] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleting the datastore file [datastore1] 8500a7e9-3fdc-411e-a48e-189d4d7bffba {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.762595] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0bce9bba-2a4c-45e3-b855-34b0bebee377 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.770963] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364191, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.773343] env[63197]: DEBUG oslo_vmware.api [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for the task: (returnval){ [ 887.773343] env[63197]: value = "task-1364196" [ 887.773343] env[63197]: _type = "Task" [ 887.773343] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.773611] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 887.773611] env[63197]: value = "task-1364195" [ 887.773611] env[63197]: _type = "Task" [ 887.773611] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.785431] env[63197]: DEBUG oslo_vmware.api [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364196, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.788988] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364195, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.837033] env[63197]: DEBUG oslo_concurrency.lockutils [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] Releasing lock "refresh_cache-1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.837323] env[63197]: DEBUG nova.compute.manager [req-aec3f101-d505-473c-b4d9-bedd0213ecd3 req-3279b1c6-a1bd-4790-adb7-718c2855bcdf service nova] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Received event network-vif-deleted-eae8fdec-5c12-41cd-b92e-8673e5d68cb0 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.923988] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48c44d1-b258-4039-8337-401e38f32150 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.931182] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ab320a-4d1c-46a7-b5b7-859f9097d985 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.961764] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f3a0a0-9cae-4de9-9e30-59477fd7be10 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.969105] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d118bad8-6fcb-4163-8d51-112e866edf68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.982861] env[63197]: DEBUG nova.compute.provider_tree [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.023598] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52059591-0ab7-0646-88b6-3acf1b7ae206, 'name': SearchDatastore_Task, 'duration_secs': 0.056911} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.023598] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.023598] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.023797] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.023931] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.024127] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.024381] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c229215e-c1a9-4737-bd88-802efd172408 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.031937] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.032134] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.032849] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d49870a4-9341-44c9-b2a0-dd72c7e2ddc0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.038147] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 888.038147] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526fb92e-36e7-69a0-f2ac-e44668d9d6a0" [ 888.038147] env[63197]: _type = "Task" [ 888.038147] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.047878] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526fb92e-36e7-69a0-f2ac-e44668d9d6a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.170804] env[63197]: DEBUG oslo_vmware.api [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264208} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.171080] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.171295] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.171440] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.171611] env[63197]: INFO nova.compute.manager [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 888.171908] env[63197]: DEBUG oslo.service.loopingcall [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.172127] env[63197]: DEBUG nova.compute.manager [-] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.172220] env[63197]: DEBUG nova.network.neutron [-] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 888.179529] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.179667] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.179819] env[63197]: DEBUG nova.network.neutron [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.227480] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364191, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58124} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.227744] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/9c80bc51-c8f6-436d-9d19-47f6d1aadbf6.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 888.231045] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 888.231045] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72c844e4-b11a-4e76-8f12-2b653566b149 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.235059] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 888.235059] env[63197]: value = "task-1364197" [ 888.235059] env[63197]: _type = "Task" [ 888.235059] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.248232] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364197, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.286048] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364195, 'name': ReconfigVM_Task, 'duration_secs': 0.309011} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.289028] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 5b9c6a1c-4212-4941-aa6c-364fb82f5e64/5b9c6a1c-4212-4941-aa6c-364fb82f5e64.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.289706] env[63197]: DEBUG oslo_vmware.api [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Task: {'id': task-1364196, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243799} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.290055] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43a35cfe-8429-47ad-838d-962102b0ce45 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.292021] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.292141] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.292361] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.292571] env[63197]: INFO nova.compute.manager [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Took 1.19 seconds to destroy the instance on the hypervisor. [ 888.292854] env[63197]: DEBUG oslo.service.loopingcall [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.293580] env[63197]: DEBUG nova.compute.manager [-] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.293580] env[63197]: DEBUG nova.network.neutron [-] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 888.300523] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 888.300523] env[63197]: value = "task-1364198" [ 888.300523] env[63197]: _type = "Task" [ 888.300523] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.310655] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364198, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.487365] env[63197]: DEBUG nova.scheduler.client.report [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.523635] env[63197]: DEBUG nova.compute.manager [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Received event network-vif-plugged-e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.523866] env[63197]: DEBUG oslo_concurrency.lockutils [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] Acquiring lock "638ef9c9-253b-4958-a660-6c1801408a51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.523932] env[63197]: DEBUG oslo_concurrency.lockutils [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] Lock "638ef9c9-253b-4958-a660-6c1801408a51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.524187] env[63197]: DEBUG oslo_concurrency.lockutils [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] Lock "638ef9c9-253b-4958-a660-6c1801408a51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.524457] env[63197]: DEBUG nova.compute.manager [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] No waiting events found dispatching network-vif-plugged-e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.524731] env[63197]: WARNING nova.compute.manager [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Received unexpected event network-vif-plugged-e927a6cd-692a-4cce-b2f6-8d321333e34f for instance with vm_state building and task_state spawning. [ 888.525014] env[63197]: DEBUG nova.compute.manager [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Received event network-changed-e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.525308] env[63197]: DEBUG nova.compute.manager [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Refreshing instance network info cache due to event network-changed-e927a6cd-692a-4cce-b2f6-8d321333e34f. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.525568] env[63197]: DEBUG oslo_concurrency.lockutils [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] Acquiring lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.551580] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526fb92e-36e7-69a0-f2ac-e44668d9d6a0, 'name': SearchDatastore_Task, 'duration_secs': 0.012175} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.552652] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec2464c1-aa0d-445c-918c-a25ee3c4af3e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.559102] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 888.559102] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5233fd9f-231d-d76f-0af9-b4d081f6241e" [ 888.559102] env[63197]: _type = "Task" [ 888.559102] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.584561] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5233fd9f-231d-d76f-0af9-b4d081f6241e, 'name': SearchDatastore_Task, 'duration_secs': 0.008622} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.584925] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.585290] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.586199] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-276ae7c6-732f-4002-bbb6-9331e2571d23 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.594637] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 888.594637] env[63197]: value = "task-1364199" [ 888.594637] env[63197]: _type = "Task" [ 888.594637] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.605733] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.710751] env[63197]: DEBUG nova.network.neutron [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.745638] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364197, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068463} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.745949] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.746770] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff217d0-4428-4d51-b0b3-16520a9305af {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.770609] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/9c80bc51-c8f6-436d-9d19-47f6d1aadbf6.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.773310] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6093033c-5d7c-44df-9ac1-f64bbfc8a3e0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.793792] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 888.793792] env[63197]: value = "task-1364200" [ 888.793792] env[63197]: _type = "Task" [ 888.793792] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.802766] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364200, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.810897] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364198, 'name': Rename_Task, 'duration_secs': 0.149961} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.810897] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.811134] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca9f514b-5aa0-46ba-a164-d6f4df6764c4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.817522] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 888.817522] env[63197]: value = "task-1364201" [ 888.817522] env[63197]: _type = "Task" [ 888.817522] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.827266] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.958242] env[63197]: DEBUG nova.network.neutron [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updating instance_info_cache with network_info: [{"id": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "address": "fa:16:3e:7e:b1:97", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape927a6cd-69", "ovs_interfaceid": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.000119] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.000119] env[63197]: DEBUG nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.001275] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.194s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.001645] env[63197]: DEBUG nova.objects.instance [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63197) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 889.074161] env[63197]: DEBUG nova.network.neutron [-] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.109486] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364199, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.156033] env[63197]: DEBUG nova.network.neutron [-] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.312662] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364200, 'name': ReconfigVM_Task, 'duration_secs': 0.470919} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.313466] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/9c80bc51-c8f6-436d-9d19-47f6d1aadbf6.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.313901] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-34a46bce-1e17-4292-8972-8cb5b2a2fd15 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.322303] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 889.322303] env[63197]: value = "task-1364202" [ 889.322303] env[63197]: _type = "Task" [ 889.322303] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.329216] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364201, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.334657] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364202, 'name': Rename_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.460921] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.460921] env[63197]: DEBUG nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Instance network_info: |[{"id": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "address": "fa:16:3e:7e:b1:97", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape927a6cd-69", "ovs_interfaceid": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.461225] env[63197]: DEBUG oslo_concurrency.lockutils [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] Acquired lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.461392] env[63197]: DEBUG nova.network.neutron [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Refreshing network info cache for port e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.462699] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:b1:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e927a6cd-692a-4cce-b2f6-8d321333e34f', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.474574] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Creating folder: Project (527721557235413e99c4a2eaa086486c). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.481357] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-60b4aa36-482d-44af-930d-26c6cd6e96f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.496500] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Created folder: Project (527721557235413e99c4a2eaa086486c) in parent group-v290286. [ 889.496500] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Creating folder: Instances. Parent ref: group-v290338. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.496500] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bf2fb889-f7b9-4a8b-b330-9cf2681b2b36 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.502925] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Created folder: Instances in parent group-v290338. [ 889.503222] env[63197]: DEBUG oslo.service.loopingcall [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.503465] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.503731] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf73792a-6196-45d5-86a2-8964be72f396 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.525279] env[63197]: DEBUG nova.compute.utils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.532705] env[63197]: DEBUG nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 889.532880] env[63197]: DEBUG nova.network.neutron [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 889.540610] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.540610] env[63197]: value = "task-1364205" [ 889.540610] env[63197]: _type = "Task" [ 889.540610] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.549511] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364205, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.576474] env[63197]: INFO nova.compute.manager [-] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Took 1.40 seconds to deallocate network for instance. [ 889.609167] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640465} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.609167] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 889.609167] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.609434] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2d0d1792-5754-4a3e-aa14-8d87a141b27b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.615387] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 889.615387] env[63197]: value = "task-1364206" [ 889.615387] env[63197]: _type = "Task" [ 889.615387] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.627680] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364206, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.651748] env[63197]: DEBUG nova.policy [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ccf6464e6854969a32f2388f7866834', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '558e5440e64c43f59d0f2cc54c18c79f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.657981] env[63197]: INFO nova.compute.manager [-] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Took 1.36 seconds to deallocate network for instance. [ 889.841959] env[63197]: DEBUG oslo_vmware.api [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364201, 'name': PowerOnVM_Task, 'duration_secs': 0.679323} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.850240] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.850240] env[63197]: INFO nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Took 9.85 seconds to spawn the instance on the hypervisor. [ 889.850240] env[63197]: DEBUG nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.850240] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364202, 'name': Rename_Task, 'duration_secs': 0.140145} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.850240] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44ad9ba3-4bea-4769-b073-385e5b883982 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.852362] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 889.852528] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d0fe198-deca-4009-ae4b-1ec72ec37d60 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.862278] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 889.862278] env[63197]: value = "task-1364207" [ 889.862278] env[63197]: _type = "Task" [ 889.862278] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.872682] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.888864] env[63197]: DEBUG nova.network.neutron [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updated VIF entry in instance network info cache for port e927a6cd-692a-4cce-b2f6-8d321333e34f. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.889276] env[63197]: DEBUG nova.network.neutron [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updating instance_info_cache with network_info: [{"id": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "address": "fa:16:3e:7e:b1:97", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape927a6cd-69", "ovs_interfaceid": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.033814] env[63197]: DEBUG nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.039596] env[63197]: DEBUG oslo_concurrency.lockutils [None req-60902504-3cee-44dc-9a28-3cd40cd0034e tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.038s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.045547] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.682s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.045845] env[63197]: INFO nova.compute.claims [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.069046] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364205, 'name': CreateVM_Task, 'duration_secs': 0.376815} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.069046] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.069550] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.069793] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.070180] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.071224] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8cd970d-b383-4705-b057-da249770846c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.076885] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 890.076885] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52455117-5eee-b3b5-d735-d956cba2a0fd" [ 890.076885] env[63197]: _type = "Task" [ 890.076885] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.083742] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.085312] env[63197]: DEBUG nova.network.neutron [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Successfully created port: 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.093127] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52455117-5eee-b3b5-d735-d956cba2a0fd, 'name': SearchDatastore_Task, 'duration_secs': 0.010412} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.093516] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.093865] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.094202] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.094524] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.094766] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.095156] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ef15881-dfde-4e31-9f53-a361a6fcca9d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.104160] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.104349] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.105196] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-133fcff1-691e-49ca-92dc-ffc26446bb64 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.111042] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 890.111042] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52140b1d-14e8-349e-752a-3436611ed9f3" [ 890.111042] env[63197]: _type = "Task" [ 890.111042] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.122431] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52140b1d-14e8-349e-752a-3436611ed9f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.127528] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364206, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105469} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.127779] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.129100] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f818d29e-adcb-41b7-8f26-cd2411603769 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.156985] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.157341] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab386b64-3997-4244-b8ad-6ce391bd63a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.173857] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.180529] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 890.180529] env[63197]: value = "task-1364208" [ 890.180529] env[63197]: _type = "Task" [ 890.180529] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.188260] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364208, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.383465] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364207, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.383988] env[63197]: INFO nova.compute.manager [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Took 25.64 seconds to build instance. [ 890.391884] env[63197]: DEBUG oslo_concurrency.lockutils [req-aac313ee-89d8-4484-bc3e-9dabaefea38f req-f6912dd8-289e-4caa-a050-3749ee98b32f service nova] Releasing lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.556736] env[63197]: DEBUG nova.compute.manager [req-74315143-885b-4cfe-ac9c-8e4706be1790 req-dedb8fab-de55-4268-9f8c-84a2e777f3a1 service nova] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Received event network-vif-deleted-c3d0ad10-09c1-4552-b5c0-5a79d8c20724 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.558912] env[63197]: DEBUG nova.compute.manager [req-74315143-885b-4cfe-ac9c-8e4706be1790 req-dedb8fab-de55-4268-9f8c-84a2e777f3a1 service nova] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Received event network-vif-deleted-619ff4d4-1f0d-4d8e-87be-1cdcf30874a5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.621266] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52140b1d-14e8-349e-752a-3436611ed9f3, 'name': SearchDatastore_Task, 'duration_secs': 0.018785} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.622245] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09fc76a2-234c-4628-8869-c4341f6807a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.628199] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 890.628199] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529c6fb2-be93-833f-d892-c4ea1478d81c" [ 890.628199] env[63197]: _type = "Task" [ 890.628199] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.640091] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529c6fb2-be93-833f-d892-c4ea1478d81c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.690224] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364208, 'name': ReconfigVM_Task, 'duration_secs': 0.277135} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.690546] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.691245] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7600312-320c-47c4-a827-7d12106ea7ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.697135] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 890.697135] env[63197]: value = "task-1364209" [ 890.697135] env[63197]: _type = "Task" [ 890.697135] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.705698] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364209, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.878740] env[63197]: DEBUG oslo_vmware.api [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364207, 'name': PowerOnVM_Task, 'duration_secs': 0.83909} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.879021] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 890.879223] env[63197]: INFO nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Took 8.91 seconds to spawn the instance on the hypervisor. [ 890.879414] env[63197]: DEBUG nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 890.883114] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c45193-1e26-4243-9500-f762082073fb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.886135] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9c420f28-b91a-4497-b08b-d7a2c30a422c tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.749s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.047960] env[63197]: DEBUG nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.091687] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.092048] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.092215] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.092512] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.092706] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.093285] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.093575] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.093777] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.094065] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.094228] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.094447] env[63197]: DEBUG nova.virt.hardware [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.095381] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d56142-25a1-416d-824d-1b8b994b44d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.106148] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a54b85-3bd2-4d7d-b784-1aa15844cd1c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.138556] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529c6fb2-be93-833f-d892-c4ea1478d81c, 'name': SearchDatastore_Task, 'duration_secs': 0.01013} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.138818] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.139080] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 638ef9c9-253b-4958-a660-6c1801408a51/638ef9c9-253b-4958-a660-6c1801408a51.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.139352] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b913685c-ce9f-4fd5-bfe1-738f0ba5b929 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.146700] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 891.146700] env[63197]: value = "task-1364210" [ 891.146700] env[63197]: _type = "Task" [ 891.146700] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.157529] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364210, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.207234] env[63197]: DEBUG nova.compute.manager [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Received event network-changed-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.207358] env[63197]: DEBUG nova.compute.manager [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Refreshing instance network info cache due to event network-changed-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 891.208139] env[63197]: DEBUG oslo_concurrency.lockutils [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] Acquiring lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.208376] env[63197]: DEBUG oslo_concurrency.lockutils [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] Acquired lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.208637] env[63197]: DEBUG nova.network.neutron [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Refreshing network info cache for port ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.219397] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364209, 'name': Rename_Task, 'duration_secs': 0.133465} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.221019] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.221019] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9421a1d8-7c8d-4ac5-a606-97b92e2e5b16 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.229094] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 891.229094] env[63197]: value = "task-1364211" [ 891.229094] env[63197]: _type = "Task" [ 891.229094] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.239843] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364211, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.327380] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a44ffa2-dea7-4d25-9258-5e0696bb7762 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.335207] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ddf43f-5f5e-46e6-90ef-4c7694846555 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.371673] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6b04ae-3b6d-4ff9-aa41-5f41bb1fd7bd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.382721] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c9bea6-6448-4b1d-84f0-ade37d883d3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.407083] env[63197]: DEBUG nova.compute.provider_tree [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.410689] env[63197]: INFO nova.compute.manager [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Took 25.51 seconds to build instance. [ 891.658071] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364210, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.708549] env[63197]: DEBUG nova.network.neutron [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Successfully updated port: 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 891.743364] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364211, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.913438] env[63197]: DEBUG nova.scheduler.client.report [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.917705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-9e29db6e-1d6d-4aac-8182-7eb7f8d5781b tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.522s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.073022] env[63197]: DEBUG nova.network.neutron [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updated VIF entry in instance network info cache for port ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.073022] env[63197]: DEBUG nova.network.neutron [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updating instance_info_cache with network_info: [{"id": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "address": "fa:16:3e:c3:5f:0c", "network": {"id": "138b731d-9b6e-45ea-b051-6a7e5c8a2471", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1660823209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.222", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "104c325de57847fd9d55ccdb10b8af97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69f65356-c85e-4b7f-ad28-7c7b5e8cf50c", "external-id": "nsx-vlan-transportzone-281", "segmentation_id": 281, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapae6d2ab9-dd", "ovs_interfaceid": "ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.158264] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364210, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538796} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.158514] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 638ef9c9-253b-4958-a660-6c1801408a51/638ef9c9-253b-4958-a660-6c1801408a51.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.158718] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.158963] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e49e888-318e-4bc5-b8b8-636d0bf61146 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.165586] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 892.165586] env[63197]: value = "task-1364212" [ 892.165586] env[63197]: _type = "Task" [ 892.165586] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.176798] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364212, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.210807] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.210960] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.211139] env[63197]: DEBUG nova.network.neutron [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.240587] env[63197]: DEBUG oslo_vmware.api [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364211, 'name': PowerOnVM_Task, 'duration_secs': 0.749471} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.240587] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.240587] env[63197]: INFO nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Took 7.86 seconds to spawn the instance on the hypervisor. [ 892.242120] env[63197]: DEBUG nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.242120] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137e6b0a-cc5f-4159-a183-0eb484e5a64e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.421527] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.422132] env[63197]: DEBUG nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.424866] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.810s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.425108] env[63197]: DEBUG nova.objects.instance [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lazy-loading 'resources' on Instance uuid 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.577386] env[63197]: DEBUG oslo_concurrency.lockutils [req-b7c4eba2-9d21-4437-8849-948cc13eac0a req-b0499ff9-e1ac-43dc-8d08-5251d6f4336f service nova] Releasing lock "refresh_cache-5b9c6a1c-4212-4941-aa6c-364fb82f5e64" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.585602] env[63197]: DEBUG nova.compute.manager [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Received event network-vif-plugged-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.585825] env[63197]: DEBUG oslo_concurrency.lockutils [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] Acquiring lock "232ea430-6a11-4dbc-9da9-0f734471d03b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.585900] env[63197]: DEBUG oslo_concurrency.lockutils [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.586076] env[63197]: DEBUG oslo_concurrency.lockutils [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.586242] env[63197]: DEBUG nova.compute.manager [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] No waiting events found dispatching network-vif-plugged-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.586401] env[63197]: WARNING nova.compute.manager [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Received unexpected event network-vif-plugged-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d for instance with vm_state building and task_state spawning. [ 892.586555] env[63197]: DEBUG nova.compute.manager [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Received event network-changed-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.586703] env[63197]: DEBUG nova.compute.manager [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Refreshing instance network info cache due to event network-changed-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 892.586861] env[63197]: DEBUG oslo_concurrency.lockutils [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] Acquiring lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.675839] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364212, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074309} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.676332] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.677225] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2967608e-fcf1-4fac-97fc-32ced60ab91c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.701915] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 638ef9c9-253b-4958-a660-6c1801408a51/638ef9c9-253b-4958-a660-6c1801408a51.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.702325] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2144a1e2-01fc-4d2e-9265-427aec85e42f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.723688] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 892.723688] env[63197]: value = "task-1364213" [ 892.723688] env[63197]: _type = "Task" [ 892.723688] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.731624] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364213, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.745038] env[63197]: DEBUG nova.network.neutron [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.768303] env[63197]: INFO nova.compute.manager [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Took 26.78 seconds to build instance. [ 892.862185] env[63197]: INFO nova.compute.manager [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Rescuing [ 892.862467] env[63197]: DEBUG oslo_concurrency.lockutils [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.862628] env[63197]: DEBUG oslo_concurrency.lockutils [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.862775] env[63197]: DEBUG nova.network.neutron [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.909887] env[63197]: DEBUG nova.network.neutron [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Updating instance_info_cache with network_info: [{"id": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "address": "fa:16:3e:f8:ef:1b", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5685fd2b-99", "ovs_interfaceid": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.929118] env[63197]: DEBUG nova.compute.utils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.933430] env[63197]: DEBUG nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.933717] env[63197]: DEBUG nova.network.neutron [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 892.989269] env[63197]: DEBUG nova.policy [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0788b47942724901a9af191269c400e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f35fa478269e4de68ae997e79237b25f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.134211] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3982fa05-22b7-423d-8b8f-45c7f9fa2b74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.148385] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405fc33f-0abf-4ff0-a035-85cac3ca1a74 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.181301] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f12b10-d963-44f2-b2a8-48948b71e880 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.188910] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5152bc-1088-4c08-9021-2bdcbdbbab13 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.202469] env[63197]: DEBUG nova.compute.provider_tree [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.233572] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364213, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.273344] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8df38a4-052d-4014-a5b2-325f05b06231 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.788s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.326688] env[63197]: DEBUG nova.network.neutron [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Successfully created port: 62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.412338] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Releasing lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.412677] env[63197]: DEBUG nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Instance network_info: |[{"id": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "address": "fa:16:3e:f8:ef:1b", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5685fd2b-99", "ovs_interfaceid": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 893.413031] env[63197]: DEBUG oslo_concurrency.lockutils [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] Acquired lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.413438] env[63197]: DEBUG nova.network.neutron [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Refreshing network info cache for port 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 893.415329] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:ef:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee4b2432-c393-4e50-ae0e-b5e12bad37db', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5685fd2b-99cb-4b1e-b61d-9ff2f290f33d', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.428934] env[63197]: DEBUG oslo.service.loopingcall [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.430640] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 893.430874] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8e594c2-f388-4e55-8aa2-34a3aa930f6d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.447314] env[63197]: DEBUG nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.455819] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.455819] env[63197]: value = "task-1364214" [ 893.455819] env[63197]: _type = "Task" [ 893.455819] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.464155] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364214, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.706299] env[63197]: DEBUG nova.scheduler.client.report [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.733578] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364213, 'name': ReconfigVM_Task, 'duration_secs': 0.701393} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.733858] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 638ef9c9-253b-4958-a660-6c1801408a51/638ef9c9-253b-4958-a660-6c1801408a51.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.734780] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-de3f9b7c-9319-437b-ab0b-21f156e17371 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.740696] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 893.740696] env[63197]: value = "task-1364215" [ 893.740696] env[63197]: _type = "Task" [ 893.740696] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.748970] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364215, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.907343] env[63197]: DEBUG nova.network.neutron [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Updating instance_info_cache with network_info: [{"id": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "address": "fa:16:3e:18:1d:83", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa42d56b7-bb", "ovs_interfaceid": "a42d56b7-bb6b-40d8-a491-2304a1e4a65b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.966636] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364214, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.145364] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 894.145575] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 894.206056] env[63197]: DEBUG nova.network.neutron [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Updated VIF entry in instance network info cache for port 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 894.206485] env[63197]: DEBUG nova.network.neutron [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Updating instance_info_cache with network_info: [{"id": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "address": "fa:16:3e:f8:ef:1b", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5685fd2b-99", "ovs_interfaceid": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.210967] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.213405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.963s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.215091] env[63197]: INFO nova.compute.claims [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.238499] env[63197]: INFO nova.scheduler.client.report [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Deleted allocations for instance 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e [ 894.250837] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364215, 'name': Rename_Task, 'duration_secs': 0.182271} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.251372] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.251610] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ebcbd5d-791b-4984-b178-65731fa0a16a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.258037] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 894.258037] env[63197]: value = "task-1364216" [ 894.258037] env[63197]: _type = "Task" [ 894.258037] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.265337] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364216, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.410042] env[63197]: DEBUG oslo_concurrency.lockutils [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "refresh_cache-9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.459171] env[63197]: DEBUG nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.471623] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364214, 'name': CreateVM_Task, 'duration_secs': 0.514063} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.471803] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.472579] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.473227] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.473227] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 894.475181] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad79b036-5c26-402c-8d81-e2516b967782 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.480565] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 894.480565] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523347d5-b4d8-5901-9a35-33b4f9288bfd" [ 894.480565] env[63197]: _type = "Task" [ 894.480565] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.486909] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.487194] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.487377] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.487568] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.487738] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.487856] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.488092] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.488276] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.488472] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.488659] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.488982] env[63197]: DEBUG nova.virt.hardware [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.489808] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6606b365-28a4-416d-a186-674aa763643e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.499305] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523347d5-b4d8-5901-9a35-33b4f9288bfd, 'name': SearchDatastore_Task, 'duration_secs': 0.014223} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.501595] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.501940] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.502127] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.502290] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.502491] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.502828] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-10ad6f9c-aa83-402f-aa97-83669afb3d22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.506516] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b17fd49-2f9e-4975-8105-a836c6a372e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.523243] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.523444] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 894.524158] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6f007b7-bddf-4bbf-9310-e94d8d11c194 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.529300] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 894.529300] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52824520-b93b-1e48-5a36-7793707df95e" [ 894.529300] env[63197]: _type = "Task" [ 894.529300] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.536720] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52824520-b93b-1e48-5a36-7793707df95e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.562835] env[63197]: INFO nova.compute.manager [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Rebuilding instance [ 894.603373] env[63197]: DEBUG nova.compute.manager [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.604380] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64629bf-1942-405f-b3a6-1524f4787b6f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.652298] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 894.652490] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 894.652629] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Rebuilding the list of instances to heal {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 894.708629] env[63197]: DEBUG oslo_concurrency.lockutils [req-6a750183-6e69-4456-9715-be01755ced8b req-69796a90-d825-4c1a-a24d-90122e6a5db2 service nova] Releasing lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.750592] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fd0e0ddc-5391-433b-9f35-cd71f96257f2 tempest-ServerShowV254Test-1410850296 tempest-ServerShowV254Test-1410850296-project-member] Lock "3b7aa0b0-d115-4026-a984-0e9fc73f2d2e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.957s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.771088] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364216, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.941250] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 894.941544] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0409ee40-0965-476c-8277-add0ede73cd4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.948989] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 894.948989] env[63197]: value = "task-1364217" [ 894.948989] env[63197]: _type = "Task" [ 894.948989] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.965952] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364217, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.040434] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52824520-b93b-1e48-5a36-7793707df95e, 'name': SearchDatastore_Task, 'duration_secs': 0.010394} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.041470] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f8798da-f2bc-4c3b-889e-2ae241810046 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.047488] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 895.047488] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52cf755f-02ad-cc87-7f79-cadd0f12ea1d" [ 895.047488] env[63197]: _type = "Task" [ 895.047488] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.055637] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52cf755f-02ad-cc87-7f79-cadd0f12ea1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.116073] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.116399] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83da75dd-266b-4714-80b4-97aab85bb61c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.123738] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 895.123738] env[63197]: value = "task-1364218" [ 895.123738] env[63197]: _type = "Task" [ 895.123738] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.136054] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.162069] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 895.162270] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 895.162595] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 895.162720] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 895.225017] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.225017] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquired lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.225017] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Forcefully refreshing network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 895.225017] env[63197]: DEBUG nova.objects.instance [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lazy-loading 'info_cache' on Instance uuid 65df32fa-a6e7-4c3a-af8b-0422df8d9229 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.274706] env[63197]: DEBUG oslo_vmware.api [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364216, 'name': PowerOnVM_Task, 'duration_secs': 0.591841} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.277503] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.277723] env[63197]: INFO nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Took 8.50 seconds to spawn the instance on the hypervisor. [ 895.277905] env[63197]: DEBUG nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.279112] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc54131-e936-4141-8c39-dc78d328b750 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.344806] env[63197]: DEBUG nova.compute.manager [req-fc38f493-3be5-4293-aa4b-415a3cf06f82 req-7b6104ad-66bb-4a60-a50c-a70929ac297d service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-vif-plugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.345571] env[63197]: DEBUG oslo_concurrency.lockutils [req-fc38f493-3be5-4293-aa4b-415a3cf06f82 req-7b6104ad-66bb-4a60-a50c-a70929ac297d service nova] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.345571] env[63197]: DEBUG oslo_concurrency.lockutils [req-fc38f493-3be5-4293-aa4b-415a3cf06f82 req-7b6104ad-66bb-4a60-a50c-a70929ac297d service nova] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.345571] env[63197]: DEBUG oslo_concurrency.lockutils [req-fc38f493-3be5-4293-aa4b-415a3cf06f82 req-7b6104ad-66bb-4a60-a50c-a70929ac297d service nova] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.345571] env[63197]: DEBUG nova.compute.manager [req-fc38f493-3be5-4293-aa4b-415a3cf06f82 req-7b6104ad-66bb-4a60-a50c-a70929ac297d service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] No waiting events found dispatching network-vif-plugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.345782] env[63197]: WARNING nova.compute.manager [req-fc38f493-3be5-4293-aa4b-415a3cf06f82 req-7b6104ad-66bb-4a60-a50c-a70929ac297d service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received unexpected event network-vif-plugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 for instance with vm_state building and task_state spawning. [ 895.406048] env[63197]: DEBUG nova.network.neutron [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Successfully updated port: 62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.458747] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364217, 'name': PowerOffVM_Task, 'duration_secs': 0.228366} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.461213] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.462215] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e6969e-f875-449d-a74d-4bcdb29b14a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.490120] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94de6628-5c38-47f4-a1a9-7737c3b7769b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.516928] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25f2f5a-97da-474a-b7c1-3a8984b44c51 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.525904] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5741f4d1-64d4-434d-9944-949af7400a19 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.531207] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 895.531747] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e1c6f83-629c-451b-801c-ffb4709f7933 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.566506] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808373d2-6fb5-41b8-b9c8-b5203b441c5e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.569055] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 895.569055] env[63197]: value = "task-1364219" [ 895.569055] env[63197]: _type = "Task" [ 895.569055] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.577940] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52cf755f-02ad-cc87-7f79-cadd0f12ea1d, 'name': SearchDatastore_Task, 'duration_secs': 0.010884} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.578669] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.579492] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 232ea430-6a11-4dbc-9da9-0f734471d03b/232ea430-6a11-4dbc-9da9-0f734471d03b.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 895.580759] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2594e83-b9f6-407f-b7f1-e94def507b82 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.587683] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4cb10bde-0f3e-4854-aff8-412eb2d4ec9c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.589902] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] VM already powered off {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 895.590173] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.590477] env[63197]: DEBUG oslo_concurrency.lockutils [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.590671] env[63197]: DEBUG oslo_concurrency.lockutils [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.590889] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.599107] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6f59828b-9cc6-4e56-a511-f3a7cc8fc7d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.601254] env[63197]: DEBUG nova.compute.provider_tree [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.604014] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 895.604014] env[63197]: value = "task-1364220" [ 895.604014] env[63197]: _type = "Task" [ 895.604014] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.609822] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.609999] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.613322] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-334e1cb4-3e46-4d55-afe3-a088789e53bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.615452] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.619404] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 895.619404] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fe083a-e0fc-071d-47f8-fe53daf65f91" [ 895.619404] env[63197]: _type = "Task" [ 895.619404] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.627046] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fe083a-e0fc-071d-47f8-fe53daf65f91, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.634109] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364218, 'name': PowerOffVM_Task, 'duration_secs': 0.308482} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.634355] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 895.634569] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 895.635292] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb6c5d90-c97e-40c0-b8b2-2ca32ee15fab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.641403] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 895.641585] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67f23f9f-8499-439c-8165-b8a0ddbac42c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.713237] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 895.713618] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 895.713618] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleting the datastore file [datastore2] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.713862] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7e991bf-01b6-4d1c-bb20-173534746bd7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.720495] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 895.720495] env[63197]: value = "task-1364222" [ 895.720495] env[63197]: _type = "Task" [ 895.720495] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.733802] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364222, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.800209] env[63197]: INFO nova.compute.manager [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Took 29.00 seconds to build instance. [ 895.907078] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.907210] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.907323] env[63197]: DEBUG nova.network.neutron [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.106370] env[63197]: DEBUG nova.scheduler.client.report [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.118371] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364220, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.132534] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fe083a-e0fc-071d-47f8-fe53daf65f91, 'name': SearchDatastore_Task, 'duration_secs': 0.02187} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.133302] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2cd5212-49ce-4f36-9580-aa19b150f65d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.138844] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 896.138844] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aa6465-239f-80c0-0fd3-6125c24cad90" [ 896.138844] env[63197]: _type = "Task" [ 896.138844] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.149042] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aa6465-239f-80c0-0fd3-6125c24cad90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.231335] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364222, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.304335] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c93ba63a-b72b-4b46-9a3b-a5f487ab0362 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.952s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.484507] env[63197]: DEBUG nova.network.neutron [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 896.614122] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.400s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.614337] env[63197]: DEBUG nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 896.616817] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.445s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.617039] env[63197]: DEBUG nova.objects.instance [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lazy-loading 'resources' on Instance uuid 72f9b0c4-69b9-49f2-8665-ff77151883af {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.624454] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.925928} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.624454] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 232ea430-6a11-4dbc-9da9-0f734471d03b/232ea430-6a11-4dbc-9da9-0f734471d03b.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.624454] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.624454] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-84fefaa7-b041-410d-b64f-7dcb88568aaa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.630579] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 896.630579] env[63197]: value = "task-1364223" [ 896.630579] env[63197]: _type = "Task" [ 896.630579] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.639274] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364223, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.647765] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aa6465-239f-80c0-0fd3-6125c24cad90, 'name': SearchDatastore_Task, 'duration_secs': 0.071579} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.648031] env[63197]: DEBUG oslo_concurrency.lockutils [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.648296] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk. {{(pid=63197) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 896.648556] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7fd2f55b-a0eb-49f0-8de6-56915a61e900 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.654501] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 896.654501] env[63197]: value = "task-1364224" [ 896.654501] env[63197]: _type = "Task" [ 896.654501] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.662233] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364224, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.731417] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364222, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.617436} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.733799] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 896.733988] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 896.734181] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 896.857402] env[63197]: DEBUG nova.network.neutron [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62e2288c-e6", "ovs_interfaceid": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.121160] env[63197]: DEBUG nova.compute.utils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.127962] env[63197]: DEBUG nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.128239] env[63197]: DEBUG nova.network.neutron [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.141339] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364223, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061799} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.141602] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 897.142451] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c71957-a11b-41bd-8d73-872512064db0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.169170] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 232ea430-6a11-4dbc-9da9-0f734471d03b/232ea430-6a11-4dbc-9da9-0f734471d03b.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.175009] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2c9eeff-85b8-46b3-a1aa-be906e685063 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.195452] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364224, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.196819] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 897.196819] env[63197]: value = "task-1364225" [ 897.196819] env[63197]: _type = "Task" [ 897.196819] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.205715] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364225, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.360406] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.360773] env[63197]: DEBUG nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Instance network_info: |[{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62e2288c-e6", "ovs_interfaceid": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.361267] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:ba:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62e2288c-e61e-467a-bd6d-1ac6fd79cb57', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.369475] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating folder: Project (f35fa478269e4de68ae997e79237b25f). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.372553] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c8783ab-193c-4817-90ae-a630bad854b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.384386] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created folder: Project (f35fa478269e4de68ae997e79237b25f) in parent group-v290286. [ 897.384386] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating folder: Instances. Parent ref: group-v290342. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.384386] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0001c302-bd5f-4fe4-b18f-ca00264effbe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.392894] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created folder: Instances in parent group-v290342. [ 897.393187] env[63197]: DEBUG oslo.service.loopingcall [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.393401] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.393621] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0147b054-2905-4308-aedc-153d12cbe887 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.420209] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.420209] env[63197]: value = "task-1364228" [ 897.420209] env[63197]: _type = "Task" [ 897.420209] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.426500] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364228, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.446289] env[63197]: DEBUG nova.policy [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f43a1f72c46efa2f36553d1a768ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d97b713fdcb4cd58c6bea06ec23a441', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.449892] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.455362] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc9c24a-9bad-4e01-8e2c-5082f7baa655 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.460716] env[63197]: DEBUG nova.compute.manager [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.460924] env[63197]: DEBUG nova.compute.manager [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing instance network info cache due to event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.461159] env[63197]: DEBUG oslo_concurrency.lockutils [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] Acquiring lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.461299] env[63197]: DEBUG oslo_concurrency.lockutils [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] Acquired lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.461764] env[63197]: DEBUG nova.network.neutron [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.468902] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8726cce-f92c-40aa-a7d6-33f14674efda {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.503585] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b05330d-616e-419d-a6d3-3c1da0f3951d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.511356] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea5d528-0919-4e23-99cd-23486c2db85f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.526852] env[63197]: DEBUG nova.compute.provider_tree [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.628717] env[63197]: DEBUG nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.670902] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364224, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.923041} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.671247] env[63197]: INFO nova.virt.vmwareapi.ds_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk. [ 897.672131] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e3320a-0d71-4343-afb1-eb60b2f37c28 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.701021] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.701784] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43ef4a3e-4c5a-41b7-869d-733d8886791e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.724221] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.725900] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 897.725900] env[63197]: value = "task-1364229" [ 897.725900] env[63197]: _type = "Task" [ 897.725900] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.734789] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364229, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.774296] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.774584] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.774744] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.774930] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.775095] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.775277] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.775495] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.775696] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.775912] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.776090] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.776268] env[63197]: DEBUG nova.virt.hardware [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.777189] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77c83e2-57d8-4884-872e-3d95ecaeb789 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.786395] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3787e0ff-3bbe-46dc-bf4b-ea2c9099bda4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.801477] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:eb:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '088f6f2e-acc0-4966-9ee5-47b6c6799c5b', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.809340] env[63197]: DEBUG oslo.service.loopingcall [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.810164] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.810164] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2c55aec-7153-4172-bde8-20ca566d34b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.833122] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.833122] env[63197]: value = "task-1364230" [ 897.833122] env[63197]: _type = "Task" [ 897.833122] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.840067] env[63197]: DEBUG nova.network.neutron [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Successfully created port: 1ad29d63-8ec8-498a-ad1d-7fd147ddc45a {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.844836] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364230, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.928416] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364228, 'name': CreateVM_Task, 'duration_secs': 0.423636} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.928621] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.929386] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.929627] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.930070] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.930370] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6545b8ff-2f85-4ac6-bf03-8ef8f6b6e831 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.935543] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 897.935543] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5260a03b-f763-ac65-37b0-38da6f66004f" [ 897.935543] env[63197]: _type = "Task" [ 897.935543] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.944892] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5260a03b-f763-ac65-37b0-38da6f66004f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.952628] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Releasing lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.952722] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updated the network info_cache for instance {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 897.953063] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 897.953280] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 897.953471] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 897.953575] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 897.953782] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 897.954253] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 897.954253] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 897.954253] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 898.030303] env[63197]: DEBUG nova.scheduler.client.report [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.216329] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364225, 'name': ReconfigVM_Task, 'duration_secs': 0.648336} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.216809] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 232ea430-6a11-4dbc-9da9-0f734471d03b/232ea430-6a11-4dbc-9da9-0f734471d03b.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.219399] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aaa0aaea-bb07-4c96-87b3-8a724a1885fa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.227017] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 898.227017] env[63197]: value = "task-1364231" [ 898.227017] env[63197]: _type = "Task" [ 898.227017] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.241357] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364231, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.241357] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364229, 'name': ReconfigVM_Task, 'duration_secs': 0.308203} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.243048] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.244023] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe846b6-f93e-4d8d-aa10-d1691ba13006 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.271230] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05d98874-720b-40d1-8c53-58a653834eee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.285706] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 898.285706] env[63197]: value = "task-1364232" [ 898.285706] env[63197]: _type = "Task" [ 898.285706] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.295233] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364232, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.342591] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364230, 'name': CreateVM_Task, 'duration_secs': 0.346546} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.342753] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.343438] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.343607] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.343931] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.344210] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d022dc5-f4f4-4f22-9088-727d7196fcc3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.348934] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 898.348934] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b77005-8485-5dee-93f2-39f6c80208e8" [ 898.348934] env[63197]: _type = "Task" [ 898.348934] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.349693] env[63197]: DEBUG nova.network.neutron [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updated VIF entry in instance network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 898.350029] env[63197]: DEBUG nova.network.neutron [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62e2288c-e6", "ovs_interfaceid": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.358912] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b77005-8485-5dee-93f2-39f6c80208e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.446717] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5260a03b-f763-ac65-37b0-38da6f66004f, 'name': SearchDatastore_Task, 'duration_secs': 0.011629} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.447029] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.447269] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.447542] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.447695] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.447875] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.448147] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90b90e48-0f2a-478d-b4b3-88a2143b9b36 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.459130] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.468494] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.468672] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.469667] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89b97de6-4754-43a8-bc72-5cfe4172026a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.474422] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 898.474422] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529c8bf2-b1cb-cd1a-b16a-aad12bbdd1d5" [ 898.474422] env[63197]: _type = "Task" [ 898.474422] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.482114] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529c8bf2-b1cb-cd1a-b16a-aad12bbdd1d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.539079] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.922s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.541412] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.468s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.543240] env[63197]: INFO nova.compute.claims [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.563607] env[63197]: INFO nova.scheduler.client.report [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Deleted allocations for instance 72f9b0c4-69b9-49f2-8665-ff77151883af [ 898.639259] env[63197]: DEBUG nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.664053] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.664322] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.664477] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.664656] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.664801] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.664945] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.665169] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.665325] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.665485] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.665641] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.665810] env[63197]: DEBUG nova.virt.hardware [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.666759] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a252e2-9c05-4e02-aa89-13191020d823 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.675033] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2274ea-7299-4354-8674-b2bd7356fd40 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.734579] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364231, 'name': Rename_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.795838] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364232, 'name': ReconfigVM_Task, 'duration_secs': 0.15227} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.796215] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.796376] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f4576a0-b293-49d3-ba63-f57897feb620 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.802507] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 898.802507] env[63197]: value = "task-1364233" [ 898.802507] env[63197]: _type = "Task" [ 898.802507] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.809913] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364233, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.855144] env[63197]: DEBUG oslo_concurrency.lockutils [req-a5ec5c11-0b13-4774-8684-ac3df6ed2dc3 req-3cacd50e-4c6a-4a54-9810-7b82983ced63 service nova] Releasing lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.861412] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b77005-8485-5dee-93f2-39f6c80208e8, 'name': SearchDatastore_Task, 'duration_secs': 0.024403} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.861713] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.862024] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.862280] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.862428] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.862601] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.863233] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-389aaf09-8865-4461-bb7e-881db4c7df9c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.875108] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.875287] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.875970] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1535c6c3-f0b6-4a46-a393-3821a6e1a599 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.881065] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 898.881065] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bf2e72-f6f7-55ca-4781-d30997d4f894" [ 898.881065] env[63197]: _type = "Task" [ 898.881065] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.887880] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bf2e72-f6f7-55ca-4781-d30997d4f894, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.984682] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529c8bf2-b1cb-cd1a-b16a-aad12bbdd1d5, 'name': SearchDatastore_Task, 'duration_secs': 0.082812} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.985660] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17fcf9cd-d8f5-45f5-bb9f-bc03b0b28598 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.991557] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 898.991557] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521500cf-c188-26d8-b04e-c8ee60775caa" [ 898.991557] env[63197]: _type = "Task" [ 898.991557] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.999107] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521500cf-c188-26d8-b04e-c8ee60775caa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.070131] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f741f6e3-15c3-465d-a90b-647812807652 tempest-ServerPasswordTestJSON-1175486418 tempest-ServerPasswordTestJSON-1175486418-project-member] Lock "72f9b0c4-69b9-49f2-8665-ff77151883af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.010s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.235941] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364231, 'name': Rename_Task, 'duration_secs': 0.528599} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.236238] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.236506] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e4fc8f5-763d-4a7a-994e-6f23f2a8f609 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.242286] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 899.242286] env[63197]: value = "task-1364234" [ 899.242286] env[63197]: _type = "Task" [ 899.242286] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.249589] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364234, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.311955] env[63197]: DEBUG oslo_vmware.api [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364233, 'name': PowerOnVM_Task, 'duration_secs': 0.389272} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.312642] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.318447] env[63197]: DEBUG nova.compute.manager [None req-df6470f8-1999-452e-b953-61b2efca4b1a tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.319268] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8373f278-a4c5-4262-8beb-0e8067375597 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.391249] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bf2e72-f6f7-55ca-4781-d30997d4f894, 'name': SearchDatastore_Task, 'duration_secs': 0.037156} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.392085] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eac73927-1127-4088-84eb-408524057a43 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.398019] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 899.398019] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d07428-bcbe-c2d3-de82-0470c5364088" [ 899.398019] env[63197]: _type = "Task" [ 899.398019] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.405234] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d07428-bcbe-c2d3-de82-0470c5364088, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.488047] env[63197]: DEBUG nova.compute.manager [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Received event network-changed-e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.488268] env[63197]: DEBUG nova.compute.manager [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Refreshing instance network info cache due to event network-changed-e927a6cd-692a-4cce-b2f6-8d321333e34f. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 899.488486] env[63197]: DEBUG oslo_concurrency.lockutils [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] Acquiring lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.488628] env[63197]: DEBUG oslo_concurrency.lockutils [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] Acquired lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.488787] env[63197]: DEBUG nova.network.neutron [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Refreshing network info cache for port e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.502805] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521500cf-c188-26d8-b04e-c8ee60775caa, 'name': SearchDatastore_Task, 'duration_secs': 0.037558} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.503103] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.503371] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.503629] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8758569f-1496-4792-bc67-c0c8c2275eda {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.510948] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 899.510948] env[63197]: value = "task-1364235" [ 899.510948] env[63197]: _type = "Task" [ 899.510948] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.519657] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364235, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.519657] env[63197]: DEBUG nova.network.neutron [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Successfully updated port: 1ad29d63-8ec8-498a-ad1d-7fd147ddc45a {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.741680] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcecc8d-b754-4103-8a84-bf3d8e02e122 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.756312] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364234, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.759387] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e999b3d-311a-46af-a412-31b660b57b2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.793208] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45927d6b-9ca0-4548-a530-d5d12c759890 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.803136] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a389fcd-d228-4be3-be25-0ed7414d6fd6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.819823] env[63197]: DEBUG nova.compute.provider_tree [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.910843] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d07428-bcbe-c2d3-de82-0470c5364088, 'name': SearchDatastore_Task, 'duration_secs': 0.010024} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.911163] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.911422] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.911702] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-209a0018-64c5-4f83-b2a5-4d7b77234b4c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.920883] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 899.920883] env[63197]: value = "task-1364236" [ 899.920883] env[63197]: _type = "Task" [ 899.920883] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.929976] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.022557] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "refresh_cache-9d8c131d-1183-4508-ae2d-28e38a50e58d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.022557] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "refresh_cache-9d8c131d-1183-4508-ae2d-28e38a50e58d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.022557] env[63197]: DEBUG nova.network.neutron [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.023542] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364235, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485829} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.026810] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 900.027037] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.027323] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82616b74-6838-44fa-9b93-8a8ea17bd51b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.037687] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 900.037687] env[63197]: value = "task-1364237" [ 900.037687] env[63197]: _type = "Task" [ 900.037687] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.050978] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364237, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.235333] env[63197]: DEBUG nova.network.neutron [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updated VIF entry in instance network info cache for port e927a6cd-692a-4cce-b2f6-8d321333e34f. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.235749] env[63197]: DEBUG nova.network.neutron [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updating instance_info_cache with network_info: [{"id": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "address": "fa:16:3e:7e:b1:97", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape927a6cd-69", "ovs_interfaceid": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.256462] env[63197]: DEBUG oslo_vmware.api [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364234, 'name': PowerOnVM_Task, 'duration_secs': 0.791038} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.260506] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.260738] env[63197]: INFO nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Took 9.21 seconds to spawn the instance on the hypervisor. [ 900.260919] env[63197]: DEBUG nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.261776] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-218133cb-4d2c-4886-a449-7a04d261acea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.348478] env[63197]: ERROR nova.scheduler.client.report [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [req-5ef44bd5-d31c-4f95-8bce-ab6df4c3d00a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e9e01824-bb93-4102-9e5b-66d5dbc1990b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5ef44bd5-d31c-4f95-8bce-ab6df4c3d00a"}]} [ 900.371480] env[63197]: DEBUG nova.scheduler.client.report [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Refreshing inventories for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 900.390526] env[63197]: DEBUG nova.scheduler.client.report [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Updating ProviderTree inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 900.390823] env[63197]: DEBUG nova.compute.provider_tree [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.406319] env[63197]: DEBUG nova.scheduler.client.report [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Refreshing aggregate associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, aggregates: None {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 900.432222] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364236, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.432222] env[63197]: DEBUG nova.scheduler.client.report [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Refreshing trait associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64 {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 900.548176] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364237, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14481} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.551062] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 900.552456] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c618df-8f43-4321-b583-9a5dde723f7d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.586522] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 900.586522] env[63197]: DEBUG nova.network.neutron [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.590348] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6fa27b5-c628-465b-8adb-a7ec06fa3cdb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.615224] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 900.615224] env[63197]: value = "task-1364238" [ 900.615224] env[63197]: _type = "Task" [ 900.615224] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.624766] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364238, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.698240] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f93531-290e-44d2-807e-e258eb01d139 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.708034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e058f4-f46f-4abb-9f5c-305a491d0b68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.741735] env[63197]: DEBUG oslo_concurrency.lockutils [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] Releasing lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.742027] env[63197]: DEBUG nova.compute.manager [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Received event network-vif-plugged-1ad29d63-8ec8-498a-ad1d-7fd147ddc45a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.742233] env[63197]: DEBUG oslo_concurrency.lockutils [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] Acquiring lock "9d8c131d-1183-4508-ae2d-28e38a50e58d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.742445] env[63197]: DEBUG oslo_concurrency.lockutils [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.742607] env[63197]: DEBUG oslo_concurrency.lockutils [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.742769] env[63197]: DEBUG nova.compute.manager [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] No waiting events found dispatching network-vif-plugged-1ad29d63-8ec8-498a-ad1d-7fd147ddc45a {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.743190] env[63197]: WARNING nova.compute.manager [req-fb7762be-1675-46a9-9121-64c3f71f32cf req-4146f1e4-8ef4-4e61-8f64-3608a958273e service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Received unexpected event network-vif-plugged-1ad29d63-8ec8-498a-ad1d-7fd147ddc45a for instance with vm_state building and task_state spawning. [ 900.744241] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee240b7-4c6b-49eb-a80a-4f7c1d4b5104 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.752665] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d861b42-b393-46e2-ba43-321c3f840844 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.767861] env[63197]: DEBUG nova.compute.provider_tree [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.787331] env[63197]: INFO nova.compute.manager [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Took 30.45 seconds to build instance. [ 900.836880] env[63197]: DEBUG nova.network.neutron [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Updating instance_info_cache with network_info: [{"id": "1ad29d63-8ec8-498a-ad1d-7fd147ddc45a", "address": "fa:16:3e:64:b5:74", "network": {"id": "ee3d27f6-c292-4851-87a3-1f79a0f1235f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1284760463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d97b713fdcb4cd58c6bea06ec23a441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ad29d63-8e", "ovs_interfaceid": "1ad29d63-8ec8-498a-ad1d-7fd147ddc45a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.933437] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364236, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.62029} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.933716] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 900.933932] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.934202] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa859766-df54-4cde-96f5-99f1a1d5bc65 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.941285] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 900.941285] env[63197]: value = "task-1364239" [ 900.941285] env[63197]: _type = "Task" [ 900.941285] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.948967] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364239, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.127739] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364238, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.274034] env[63197]: DEBUG nova.scheduler.client.report [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.288505] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0e192d31-537b-49da-8e1d-c416455c3e28 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.224s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.340196] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "refresh_cache-9d8c131d-1183-4508-ae2d-28e38a50e58d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.340409] env[63197]: DEBUG nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Instance network_info: |[{"id": "1ad29d63-8ec8-498a-ad1d-7fd147ddc45a", "address": "fa:16:3e:64:b5:74", "network": {"id": "ee3d27f6-c292-4851-87a3-1f79a0f1235f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1284760463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d97b713fdcb4cd58c6bea06ec23a441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ad29d63-8e", "ovs_interfaceid": "1ad29d63-8ec8-498a-ad1d-7fd147ddc45a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.340798] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:b5:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '447ff42d-b33e-4b5d-8b7f-e8117ebbbc92', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1ad29d63-8ec8-498a-ad1d-7fd147ddc45a', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.349407] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Creating folder: Project (5d97b713fdcb4cd58c6bea06ec23a441). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.349744] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-31d97671-6e11-4538-970b-8705e203e1ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.361012] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Created folder: Project (5d97b713fdcb4cd58c6bea06ec23a441) in parent group-v290286. [ 901.361233] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Creating folder: Instances. Parent ref: group-v290346. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.361526] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85b57cfe-5d5d-46e5-a624-082dd08682e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.372088] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Created folder: Instances in parent group-v290346. [ 901.372652] env[63197]: DEBUG oslo.service.loopingcall [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.373245] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.373612] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-439c6778-b8d6-4e8e-9cb8-0e19c54572ec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.394907] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.394907] env[63197]: value = "task-1364242" [ 901.394907] env[63197]: _type = "Task" [ 901.394907] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.405456] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364242, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.452926] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364239, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068414} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.453394] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.454113] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbbf117-99b7-4f18-9ecf-067e0d842e61 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.476718] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.477450] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36661cfe-6708-474d-b3be-0a41b31ce64f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.501065] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 901.501065] env[63197]: value = "task-1364243" [ 901.501065] env[63197]: _type = "Task" [ 901.501065] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.507186] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364243, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.523091] env[63197]: DEBUG nova.compute.manager [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Received event network-changed-1ad29d63-8ec8-498a-ad1d-7fd147ddc45a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.523363] env[63197]: DEBUG nova.compute.manager [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Refreshing instance network info cache due to event network-changed-1ad29d63-8ec8-498a-ad1d-7fd147ddc45a. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.523601] env[63197]: DEBUG oslo_concurrency.lockutils [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] Acquiring lock "refresh_cache-9d8c131d-1183-4508-ae2d-28e38a50e58d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.523817] env[63197]: DEBUG oslo_concurrency.lockutils [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] Acquired lock "refresh_cache-9d8c131d-1183-4508-ae2d-28e38a50e58d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.525702] env[63197]: DEBUG nova.network.neutron [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Refreshing network info cache for port 1ad29d63-8ec8-498a-ad1d-7fd147ddc45a {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.567022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "52549788-b58f-4f5d-8443-a0c771099884" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.567022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "52549788-b58f-4f5d-8443-a0c771099884" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.627846] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364238, 'name': ReconfigVM_Task, 'duration_secs': 0.793135} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.628214] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 901.629261] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-62597287-7adb-4f44-bb4b-b2e980a250bd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.636528] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 901.636528] env[63197]: value = "task-1364244" [ 901.636528] env[63197]: _type = "Task" [ 901.636528] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.646538] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364244, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.778019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.236s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.778632] env[63197]: DEBUG nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 901.783626] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.515s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.783869] env[63197]: DEBUG nova.objects.instance [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lazy-loading 'resources' on Instance uuid a0a40337-d8d7-448b-afff-f6849e9d37a1 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.906334] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364242, 'name': CreateVM_Task, 'duration_secs': 0.42624} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.906739] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.907181] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.907370] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.907943] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.907943] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a343fc0e-c3c9-4f30-adf2-f18d05ce9d95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.912577] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 901.912577] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5292e570-95f4-bdfb-5c71-ecfb6fe92b58" [ 901.912577] env[63197]: _type = "Task" [ 901.912577] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.920638] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5292e570-95f4-bdfb-5c71-ecfb6fe92b58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.013134] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364243, 'name': ReconfigVM_Task, 'duration_secs': 0.311201} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.013444] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d/1067806f-7a4d-4e3f-86c5-7b0d97f33d2d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.014075] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-392077f1-ab8f-4524-a16c-857a0b81c5f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.021684] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 902.021684] env[63197]: value = "task-1364245" [ 902.021684] env[63197]: _type = "Task" [ 902.021684] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.031259] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364245, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.067178] env[63197]: DEBUG nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.149111] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364244, 'name': Rename_Task, 'duration_secs': 0.192853} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.149111] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.149111] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c04c5de1-7531-4864-a9de-960438871c89 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.154959] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 902.154959] env[63197]: value = "task-1364246" [ 902.154959] env[63197]: _type = "Task" [ 902.154959] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.165820] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364246, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.255598] env[63197]: DEBUG nova.compute.manager [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.255789] env[63197]: DEBUG nova.compute.manager [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing instance network info cache due to event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.255999] env[63197]: DEBUG oslo_concurrency.lockutils [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] Acquiring lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.256158] env[63197]: DEBUG oslo_concurrency.lockutils [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] Acquired lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.256315] env[63197]: DEBUG nova.network.neutron [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.287340] env[63197]: DEBUG nova.compute.utils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.288953] env[63197]: DEBUG nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 902.289233] env[63197]: DEBUG nova.network.neutron [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 902.353378] env[63197]: DEBUG nova.network.neutron [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Updated VIF entry in instance network info cache for port 1ad29d63-8ec8-498a-ad1d-7fd147ddc45a. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.353683] env[63197]: DEBUG nova.network.neutron [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Updating instance_info_cache with network_info: [{"id": "1ad29d63-8ec8-498a-ad1d-7fd147ddc45a", "address": "fa:16:3e:64:b5:74", "network": {"id": "ee3d27f6-c292-4851-87a3-1f79a0f1235f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1284760463-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d97b713fdcb4cd58c6bea06ec23a441", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ad29d63-8e", "ovs_interfaceid": "1ad29d63-8ec8-498a-ad1d-7fd147ddc45a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.378398] env[63197]: DEBUG nova.policy [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12d00044b65c48f484ac9fea595d6179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80138785cfdb4a2188ffb972da437646', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.423401] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5292e570-95f4-bdfb-5c71-ecfb6fe92b58, 'name': SearchDatastore_Task, 'duration_secs': 0.014776} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.423574] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.423831] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.424110] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.424297] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.424506] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.424784] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-be26cf40-e9ce-4283-8514-70fd3ba873f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.441065] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.441065] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.441500] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ea70fd6-3a4c-4931-ad75-8f3193baa0c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.449423] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 902.449423] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e43e66-0f4b-6fd6-83fb-ebbaefcb481e" [ 902.449423] env[63197]: _type = "Task" [ 902.449423] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.458483] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e43e66-0f4b-6fd6-83fb-ebbaefcb481e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.532392] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364245, 'name': Rename_Task, 'duration_secs': 0.164302} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.535176] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.535623] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51afac2d-005d-4fd2-927d-09ed39521f88 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.542319] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 902.542319] env[63197]: value = "task-1364247" [ 902.542319] env[63197]: _type = "Task" [ 902.542319] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.556369] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364247, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.565353] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb32bbc-db00-4515-a5b4-30b5c442b663 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.579157] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909fecf9-f88d-4d07-9d0c-d397e172b107 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.616631] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.616631] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd20fde4-b7fc-4cf9-88f3-12f5cb492b68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.625951] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f0bafd-f5c1-4198-a3cc-f851b4200471 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.641021] env[63197]: DEBUG nova.compute.provider_tree [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.668168] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364246, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.717306] env[63197]: DEBUG nova.network.neutron [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Successfully created port: 5a2fa3ff-121f-40d3-b800-77904777a6d1 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.797530] env[63197]: DEBUG nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 902.858382] env[63197]: DEBUG oslo_concurrency.lockutils [req-503b338b-32ff-4483-95ee-61a0765ca029 req-7485a89e-943a-435b-b10c-83a090002f3f service nova] Releasing lock "refresh_cache-9d8c131d-1183-4508-ae2d-28e38a50e58d" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.964047] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e43e66-0f4b-6fd6-83fb-ebbaefcb481e, 'name': SearchDatastore_Task, 'duration_secs': 0.022555} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.965642] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ede5225b-9d3a-4db6-a11e-92d07649bf90 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.971615] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 902.971615] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d4697f-d66a-87d4-7c02-6f4dc578a08e" [ 902.971615] env[63197]: _type = "Task" [ 902.971615] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.980341] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d4697f-d66a-87d4-7c02-6f4dc578a08e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.053574] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364247, 'name': PowerOnVM_Task} progress is 90%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.146083] env[63197]: DEBUG nova.scheduler.client.report [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.172316] env[63197]: DEBUG oslo_vmware.api [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364246, 'name': PowerOnVM_Task, 'duration_secs': 0.787332} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.172316] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.172316] env[63197]: INFO nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Took 8.71 seconds to spawn the instance on the hypervisor. [ 903.172316] env[63197]: DEBUG nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.172316] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0134ea2b-e25a-4048-8253-5b0898cebb29 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.311707] env[63197]: DEBUG nova.network.neutron [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updated VIF entry in instance network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.311840] env[63197]: DEBUG nova.network.neutron [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.483172] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d4697f-d66a-87d4-7c02-6f4dc578a08e, 'name': SearchDatastore_Task, 'duration_secs': 0.012164} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.483465] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.483765] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 9d8c131d-1183-4508-ae2d-28e38a50e58d/9d8c131d-1183-4508-ae2d-28e38a50e58d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.484043] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2eeafc48-6431-4ecd-a8e1-fbbc53755bc0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.490424] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 903.490424] env[63197]: value = "task-1364248" [ 903.490424] env[63197]: _type = "Task" [ 903.490424] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.498273] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364248, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.551948] env[63197]: DEBUG oslo_vmware.api [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364247, 'name': PowerOnVM_Task, 'duration_secs': 0.623161} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.552230] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.552484] env[63197]: DEBUG nova.compute.manager [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.553318] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf64841-689e-4395-9ec3-5d67fd52ce38 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.652275] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.654509] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.150s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.657011] env[63197]: INFO nova.compute.claims [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.680783] env[63197]: INFO nova.scheduler.client.report [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Deleted allocations for instance a0a40337-d8d7-448b-afff-f6849e9d37a1 [ 903.697697] env[63197]: INFO nova.compute.manager [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Took 27.36 seconds to build instance. [ 903.814971] env[63197]: DEBUG nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 903.817372] env[63197]: DEBUG oslo_concurrency.lockutils [req-db74b0fc-0b56-44a7-8148-35faae547eda req-28bca33a-fbd2-4dcb-bc6f-1d02cd2dba43 service nova] Releasing lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.843578] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.843901] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.844129] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.844415] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.844655] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.844898] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.845272] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.845532] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.845786] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.846025] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.846270] env[63197]: DEBUG nova.virt.hardware [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.847489] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfeeeea0-cec7-40e1-b1e8-d9245dfdd51c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.857330] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a604006-0fbf-4f9f-aefc-ee7349d74a8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.999740] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364248, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.074803] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.200857] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b770a920-c388-48a3-b82f-a4f55f15a267 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.333s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.202056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5fa6506c-21d0-4b48-9790-122df0283134 tempest-InstanceActionsNegativeTestJSON-1705685655 tempest-InstanceActionsNegativeTestJSON-1705685655-project-member] Lock "a0a40337-d8d7-448b-afff-f6849e9d37a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.942s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.292731] env[63197]: DEBUG nova.compute.manager [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.293205] env[63197]: DEBUG nova.compute.manager [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing instance network info cache due to event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.299031] env[63197]: DEBUG oslo_concurrency.lockutils [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] Acquiring lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.299031] env[63197]: DEBUG oslo_concurrency.lockutils [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] Acquired lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.299031] env[63197]: DEBUG nova.network.neutron [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.504232] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364248, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.901388} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.504232] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 9d8c131d-1183-4508-ae2d-28e38a50e58d/9d8c131d-1183-4508-ae2d-28e38a50e58d.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.504232] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.504232] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76a6d9f5-711d-4495-979c-f4d05ef71a5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.509849] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 904.509849] env[63197]: value = "task-1364249" [ 904.509849] env[63197]: _type = "Task" [ 904.509849] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.519584] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364249, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.903427] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4794e9-9a86-44c9-92ec-639baa9afe53 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.912157] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d02ea48-8aa7-4f8e-9772-0dc6698d8a78 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.918678] env[63197]: DEBUG nova.compute.manager [req-e5535f3f-bfe7-41d3-8364-03786a94cf12 req-561fa84b-69f8-4903-96f5-c6cfafeb6ec8 service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Received event network-vif-plugged-5a2fa3ff-121f-40d3-b800-77904777a6d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.919098] env[63197]: DEBUG oslo_concurrency.lockutils [req-e5535f3f-bfe7-41d3-8364-03786a94cf12 req-561fa84b-69f8-4903-96f5-c6cfafeb6ec8 service nova] Acquiring lock "a1f365de-b634-481a-b2e8-9bfc801a47f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.919441] env[63197]: DEBUG oslo_concurrency.lockutils [req-e5535f3f-bfe7-41d3-8364-03786a94cf12 req-561fa84b-69f8-4903-96f5-c6cfafeb6ec8 service nova] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.919730] env[63197]: DEBUG oslo_concurrency.lockutils [req-e5535f3f-bfe7-41d3-8364-03786a94cf12 req-561fa84b-69f8-4903-96f5-c6cfafeb6ec8 service nova] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.920041] env[63197]: DEBUG nova.compute.manager [req-e5535f3f-bfe7-41d3-8364-03786a94cf12 req-561fa84b-69f8-4903-96f5-c6cfafeb6ec8 service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] No waiting events found dispatching network-vif-plugged-5a2fa3ff-121f-40d3-b800-77904777a6d1 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.920345] env[63197]: WARNING nova.compute.manager [req-e5535f3f-bfe7-41d3-8364-03786a94cf12 req-561fa84b-69f8-4903-96f5-c6cfafeb6ec8 service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Received unexpected event network-vif-plugged-5a2fa3ff-121f-40d3-b800-77904777a6d1 for instance with vm_state building and task_state spawning. [ 904.949537] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a3b73d-b549-4312-a626-0dd8e7b1082b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.962019] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f852e030-bd8f-46e9-99a5-60983f566233 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.974196] env[63197]: DEBUG nova.compute.provider_tree [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.020845] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364249, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.25734} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.021504] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.024015] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c249d5-382b-40cf-8b39-340b5feff5e1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.046082] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 9d8c131d-1183-4508-ae2d-28e38a50e58d/9d8c131d-1183-4508-ae2d-28e38a50e58d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.046576] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb833e95-557b-440e-b5bc-8e4d2797be70 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.069535] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 905.069535] env[63197]: value = "task-1364250" [ 905.069535] env[63197]: _type = "Task" [ 905.069535] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.079204] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364250, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.481020] env[63197]: DEBUG nova.scheduler.client.report [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.580774] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364250, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.629542] env[63197]: DEBUG nova.network.neutron [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updated VIF entry in instance network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 905.629939] env[63197]: DEBUG nova.network.neutron [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.691162] env[63197]: DEBUG nova.network.neutron [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Successfully updated port: 5a2fa3ff-121f-40d3-b800-77904777a6d1 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.743863] env[63197]: DEBUG nova.compute.manager [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.744079] env[63197]: DEBUG nova.compute.manager [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing instance network info cache due to event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.744349] env[63197]: DEBUG oslo_concurrency.lockutils [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] Acquiring lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.744553] env[63197]: DEBUG oslo_concurrency.lockutils [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] Acquired lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.744807] env[63197]: DEBUG nova.network.neutron [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.992529] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.993270] env[63197]: DEBUG nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 905.997603] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.914s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.997603] env[63197]: DEBUG nova.objects.instance [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lazy-loading 'resources' on Instance uuid 46fd7dec-588e-430e-b51f-9b61d9b148b6 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.080956] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364250, 'name': ReconfigVM_Task, 'duration_secs': 0.727721} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.081272] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 9d8c131d-1183-4508-ae2d-28e38a50e58d/9d8c131d-1183-4508-ae2d-28e38a50e58d.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.082473] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79a0f128-ecdf-4629-8187-6b07ad9d5db8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.089484] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.089484] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.089576] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.089836] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.090123] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.092016] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 906.092016] env[63197]: value = "task-1364251" [ 906.092016] env[63197]: _type = "Task" [ 906.092016] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.096112] env[63197]: INFO nova.compute.manager [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Terminating instance [ 906.098197] env[63197]: DEBUG nova.compute.manager [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.098488] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.099504] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fca9fb6-9298-4a7e-aa56-82c85f66f676 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.108354] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364251, 'name': Rename_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.110095] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.110597] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11a1b417-c8aa-42f2-bb13-613e84103cf9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.116532] env[63197]: DEBUG oslo_vmware.api [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 906.116532] env[63197]: value = "task-1364252" [ 906.116532] env[63197]: _type = "Task" [ 906.116532] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.126363] env[63197]: DEBUG oslo_vmware.api [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.136038] env[63197]: DEBUG oslo_concurrency.lockutils [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] Releasing lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.136038] env[63197]: DEBUG nova.compute.manager [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Received event network-changed-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.136038] env[63197]: DEBUG nova.compute.manager [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Refreshing instance network info cache due to event network-changed-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.136038] env[63197]: DEBUG oslo_concurrency.lockutils [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] Acquiring lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.136038] env[63197]: DEBUG oslo_concurrency.lockutils [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] Acquired lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.136439] env[63197]: DEBUG nova.network.neutron [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Refreshing network info cache for port 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.198038] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-a1f365de-b634-481a-b2e8-9bfc801a47f5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.198038] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-a1f365de-b634-481a-b2e8-9bfc801a47f5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.198038] env[63197]: DEBUG nova.network.neutron [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 906.212938] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "232ea430-6a11-4dbc-9da9-0f734471d03b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.213979] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.213979] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "232ea430-6a11-4dbc-9da9-0f734471d03b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.213979] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.213979] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.215974] env[63197]: INFO nova.compute.manager [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Terminating instance [ 906.221494] env[63197]: DEBUG nova.compute.manager [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.221494] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.222860] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4b6a22-e28a-4b40-97ab-add9f170754d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.231169] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.231169] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c0a34fd-f8f9-49f2-8809-92dff07a4125 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.243908] env[63197]: DEBUG oslo_vmware.api [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 906.243908] env[63197]: value = "task-1364253" [ 906.243908] env[63197]: _type = "Task" [ 906.243908] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.255874] env[63197]: DEBUG oslo_vmware.api [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.329011] env[63197]: DEBUG nova.compute.manager [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Received event network-changed-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.329437] env[63197]: DEBUG nova.compute.manager [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Refreshing instance network info cache due to event network-changed-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.329749] env[63197]: DEBUG oslo_concurrency.lockutils [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] Acquiring lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.502901] env[63197]: DEBUG nova.compute.utils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.506379] env[63197]: DEBUG nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 906.506632] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 906.551857] env[63197]: DEBUG nova.network.neutron [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updated VIF entry in instance network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.552200] env[63197]: DEBUG nova.network.neutron [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62e2288c-e6", "ovs_interfaceid": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.586601] env[63197]: DEBUG nova.policy [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91bab7add6374ce4adce53b2595dc5a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4f082d2f82c42218116efa35aee1f56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 906.605196] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364251, 'name': Rename_Task, 'duration_secs': 0.194133} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.605512] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.606112] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a36029b2-5c93-4119-a17c-ce122184fd14 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.612613] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 906.612613] env[63197]: value = "task-1364254" [ 906.612613] env[63197]: _type = "Task" [ 906.612613] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.627322] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364254, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.632516] env[63197]: DEBUG oslo_vmware.api [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364252, 'name': PowerOffVM_Task, 'duration_secs': 0.259236} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.632980] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.633332] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.633734] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f69c52e-ef9e-416a-91dd-038c041e9b98 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.704568] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.704816] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.705030] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleting the datastore file [datastore1] 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.705715] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccfd1799-4d25-4dc1-925b-a7ee8929d64e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.714060] env[63197]: DEBUG oslo_vmware.api [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 906.714060] env[63197]: value = "task-1364256" [ 906.714060] env[63197]: _type = "Task" [ 906.714060] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.731191] env[63197]: DEBUG oslo_vmware.api [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364256, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.763094] env[63197]: DEBUG oslo_vmware.api [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364253, 'name': PowerOffVM_Task, 'duration_secs': 0.351274} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.766523] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.766824] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.767227] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-befe11b3-05a2-491b-b94b-bc2511879716 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.794608] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc4df32-a7b4-46dd-8dfd-f3f4f799322b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.807658] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c52dedb-e203-459b-b3bf-2d021ddc1e3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.853095] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf66296a-8e96-4428-9a52-f2c3535300fa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.858174] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.858416] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.858579] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Deleting the datastore file [datastore2] 232ea430-6a11-4dbc-9da9-0f734471d03b {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.859558] env[63197]: DEBUG nova.network.neutron [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 906.861182] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.861819] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.863339] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5813da2e-0f4b-4b29-bde5-f59658223067 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.874034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4040e3c-6aef-469e-abf9-2859471a9f3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.878589] env[63197]: DEBUG oslo_vmware.api [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 906.878589] env[63197]: value = "task-1364258" [ 906.878589] env[63197]: _type = "Task" [ 906.878589] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.892825] env[63197]: DEBUG nova.compute.provider_tree [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.899607] env[63197]: DEBUG oslo_vmware.api [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.011653] env[63197]: DEBUG nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 907.058203] env[63197]: DEBUG oslo_concurrency.lockutils [req-6e6bc232-6222-40d3-a7da-c61f60bfde58 req-d3201934-fe22-475f-a4d6-67ada9771210 service nova] Releasing lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.093064] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Successfully created port: 4403e435-3845-48ec-a898-a1a932fea8d3 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.099830] env[63197]: DEBUG nova.network.neutron [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Updated VIF entry in instance network info cache for port 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 907.100197] env[63197]: DEBUG nova.network.neutron [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Updating instance_info_cache with network_info: [{"id": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "address": "fa:16:3e:f8:ef:1b", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5685fd2b-99", "ovs_interfaceid": "5685fd2b-99cb-4b1e-b61d-9ff2f290f33d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.123061] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364254, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.219314] env[63197]: DEBUG nova.network.neutron [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Updating instance_info_cache with network_info: [{"id": "5a2fa3ff-121f-40d3-b800-77904777a6d1", "address": "fa:16:3e:91:83:39", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a2fa3ff-12", "ovs_interfaceid": "5a2fa3ff-121f-40d3-b800-77904777a6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.225823] env[63197]: DEBUG oslo_vmware.api [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364256, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217879} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.226251] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.226424] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.226641] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.226904] env[63197]: INFO nova.compute.manager [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 907.227197] env[63197]: DEBUG oslo.service.loopingcall [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.227466] env[63197]: DEBUG nova.compute.manager [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.227606] env[63197]: DEBUG nova.network.neutron [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.367617] env[63197]: DEBUG nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 907.394750] env[63197]: DEBUG oslo_vmware.api [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.473568} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.394750] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.394750] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.394929] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.395140] env[63197]: INFO nova.compute.manager [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 907.395391] env[63197]: DEBUG oslo.service.loopingcall [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.399019] env[63197]: DEBUG nova.scheduler.client.report [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.399745] env[63197]: DEBUG nova.compute.manager [-] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.399745] env[63197]: DEBUG nova.network.neutron [-] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.478878] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Successfully created port: cf42dede-8506-413b-a119-cbe31f3be836 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.602477] env[63197]: DEBUG oslo_concurrency.lockutils [req-07a74fa3-4771-4f5d-b240-36628eefb306 req-5afdc92c-767f-45c3-b0a3-c74ef4780fcf service nova] Releasing lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.603067] env[63197]: DEBUG oslo_concurrency.lockutils [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] Acquired lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.604364] env[63197]: DEBUG nova.network.neutron [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Refreshing network info cache for port 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.624855] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364254, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.725499] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-a1f365de-b634-481a-b2e8-9bfc801a47f5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.725499] env[63197]: DEBUG nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Instance network_info: |[{"id": "5a2fa3ff-121f-40d3-b800-77904777a6d1", "address": "fa:16:3e:91:83:39", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a2fa3ff-12", "ovs_interfaceid": "5a2fa3ff-121f-40d3-b800-77904777a6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 907.725847] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:91:83:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a2fa3ff-121f-40d3-b800-77904777a6d1', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 907.734931] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Creating folder: Project (80138785cfdb4a2188ffb972da437646). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.738179] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-abd174a1-2fe8-4e8e-9ddc-217813027988 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.749042] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Created folder: Project (80138785cfdb4a2188ffb972da437646) in parent group-v290286. [ 907.749216] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Creating folder: Instances. Parent ref: group-v290349. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 907.749802] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-147ba3f6-d67f-43fb-ba17-a84f49e864b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.762198] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Created folder: Instances in parent group-v290349. [ 907.762198] env[63197]: DEBUG oslo.service.loopingcall [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.762396] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 907.763040] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7531e8a-e461-449b-9359-acec73174d1d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.783763] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 907.783763] env[63197]: value = "task-1364261" [ 907.783763] env[63197]: _type = "Task" [ 907.783763] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.794540] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364261, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.903201] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.906596] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.907126] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.733s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.907811] env[63197]: DEBUG nova.objects.instance [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lazy-loading 'resources' on Instance uuid 8500a7e9-3fdc-411e-a48e-189d4d7bffba {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.912409] env[63197]: DEBUG nova.compute.manager [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Received event network-changed-5a2fa3ff-121f-40d3-b800-77904777a6d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.912409] env[63197]: DEBUG nova.compute.manager [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Refreshing instance network info cache due to event network-changed-5a2fa3ff-121f-40d3-b800-77904777a6d1. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 907.912409] env[63197]: DEBUG oslo_concurrency.lockutils [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] Acquiring lock "refresh_cache-a1f365de-b634-481a-b2e8-9bfc801a47f5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.912409] env[63197]: DEBUG oslo_concurrency.lockutils [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] Acquired lock "refresh_cache-a1f365de-b634-481a-b2e8-9bfc801a47f5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.912409] env[63197]: DEBUG nova.network.neutron [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Refreshing network info cache for port 5a2fa3ff-121f-40d3-b800-77904777a6d1 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 907.930177] env[63197]: INFO nova.scheduler.client.report [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleted allocations for instance 46fd7dec-588e-430e-b51f-9b61d9b148b6 [ 908.020845] env[63197]: DEBUG nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 908.039656] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Successfully created port: c856ee9c-0a79-46b8-a8ec-f6283349a05e {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.050052] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.050319] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.050474] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.050716] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.050917] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.051112] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.051328] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.051484] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.051646] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.051803] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.052045] env[63197]: DEBUG nova.virt.hardware [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.052959] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb2e8cf-5cb4-4302-ad7c-09d67d95e7c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.062865] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038f1fa5-74b1-45df-a623-f0b311933c15 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.122562] env[63197]: DEBUG oslo_vmware.api [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364254, 'name': PowerOnVM_Task, 'duration_secs': 1.116672} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.122857] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.123086] env[63197]: INFO nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Took 9.48 seconds to spawn the instance on the hypervisor. [ 908.123324] env[63197]: DEBUG nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.124135] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8baf519-7ec9-4de4-99b2-69c21cb03258 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.143400] env[63197]: INFO nova.network.neutron [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Port 5685fd2b-99cb-4b1e-b61d-9ff2f290f33d from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 908.143635] env[63197]: DEBUG nova.network.neutron [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.242647] env[63197]: DEBUG nova.network.neutron [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.301045] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364261, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.434243] env[63197]: DEBUG nova.network.neutron [-] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.436713] env[63197]: DEBUG nova.compute.manager [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.437069] env[63197]: DEBUG nova.compute.manager [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing instance network info cache due to event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.437226] env[63197]: DEBUG oslo_concurrency.lockutils [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] Acquiring lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.437386] env[63197]: DEBUG oslo_concurrency.lockutils [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] Acquired lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.437564] env[63197]: DEBUG nova.network.neutron [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.445436] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6d4ef95a-f112-42f4-a931-ecadd55bb38f tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "46fd7dec-588e-430e-b51f-9b61d9b148b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.417s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.645484] env[63197]: INFO nova.compute.manager [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Took 27.41 seconds to build instance. [ 908.646558] env[63197]: DEBUG oslo_concurrency.lockutils [req-8baad4e1-bfc8-4bfb-97ea-8f48e4631394 req-a1ce4662-035e-41b3-850b-c60df33f1f3b service nova] Releasing lock "refresh_cache-232ea430-6a11-4dbc-9da9-0f734471d03b" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.687341] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7991104b-4558-4eef-b5b0-d758781f9a39 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.692303] env[63197]: DEBUG nova.network.neutron [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Updated VIF entry in instance network info cache for port 5a2fa3ff-121f-40d3-b800-77904777a6d1. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.692791] env[63197]: DEBUG nova.network.neutron [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Updating instance_info_cache with network_info: [{"id": "5a2fa3ff-121f-40d3-b800-77904777a6d1", "address": "fa:16:3e:91:83:39", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a2fa3ff-12", "ovs_interfaceid": "5a2fa3ff-121f-40d3-b800-77904777a6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.699215] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44af06b8-7892-4d49-bc97-45275a9aba46 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.732914] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188ca762-91de-4ef8-8790-d2016989d501 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.741432] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd63525d-a361-4e30-88a3-f84781db0c22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.745949] env[63197]: INFO nova.compute.manager [-] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Took 1.52 seconds to deallocate network for instance. [ 908.756949] env[63197]: DEBUG nova.compute.provider_tree [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.800273] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364261, 'name': CreateVM_Task, 'duration_secs': 0.543768} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.800389] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.801156] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.801456] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.801763] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.802095] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e170412a-1008-48ab-8a45-e8be627ec6a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.807875] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 908.807875] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526a8b2c-49e2-92f3-f848-da5f64f2dc41" [ 908.807875] env[63197]: _type = "Task" [ 908.807875] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.817747] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526a8b2c-49e2-92f3-f848-da5f64f2dc41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.942636] env[63197]: INFO nova.compute.manager [-] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Took 1.54 seconds to deallocate network for instance. [ 909.151684] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a9d3e080-5fc1-4220-8d52-51c06ed46316 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.549s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.197362] env[63197]: DEBUG oslo_concurrency.lockutils [req-baa5e979-8abe-4a3f-8d5f-9536fa23bf3b req-3c3bdbb4-38bd-4fb1-a638-c4e6f1dbb0af service nova] Releasing lock "refresh_cache-a1f365de-b634-481a-b2e8-9bfc801a47f5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.263761] env[63197]: DEBUG nova.scheduler.client.report [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.268864] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.319444] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526a8b2c-49e2-92f3-f848-da5f64f2dc41, 'name': SearchDatastore_Task, 'duration_secs': 0.010067} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.319750] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.320018] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.320278] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.320422] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.320602] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.320863] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f42eac74-fc13-4fce-9f23-cfbc9f862e82 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.334014] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.334207] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.334944] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a807644-fbc7-4388-9c64-74004e605e17 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.342019] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 909.342019] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b4c63d-5a69-2b76-698c-59078ab40258" [ 909.342019] env[63197]: _type = "Task" [ 909.342019] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.348362] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b4c63d-5a69-2b76-698c-59078ab40258, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.403133] env[63197]: DEBUG nova.network.neutron [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updated VIF entry in instance network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.403133] env[63197]: DEBUG nova.network.neutron [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.458385] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.704188] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Successfully updated port: 4403e435-3845-48ec-a898-a1a932fea8d3 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.768503] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.771647] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.312s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.771980] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.772331] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 909.772576] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.158s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.774312] env[63197]: INFO nova.compute.claims [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.777748] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67dbfc82-bf36-4b8c-8d0f-13a50f033de6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.787908] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ab3f99-40b9-4b35-be0e-903e21d486fb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.804921] env[63197]: INFO nova.scheduler.client.report [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Deleted allocations for instance 8500a7e9-3fdc-411e-a48e-189d4d7bffba [ 909.808756] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b20c984-6cdc-4742-bf66-e614768b3eff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.816251] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286e9cb6-2341-4ddc-b823-dfafbfdfdc30 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.848260] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181214MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 909.848411] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.857748] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b4c63d-5a69-2b76-698c-59078ab40258, 'name': SearchDatastore_Task, 'duration_secs': 0.04621} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.858637] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0937e0d6-a1dd-4036-8aec-fc57fae219d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.864187] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 909.864187] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52203c96-aa0d-cce4-e0a4-4118cfad2437" [ 909.864187] env[63197]: _type = "Task" [ 909.864187] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.871136] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52203c96-aa0d-cce4-e0a4-4118cfad2437, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.907685] env[63197]: DEBUG oslo_concurrency.lockutils [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] Releasing lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.908130] env[63197]: DEBUG nova.compute.manager [req-820d3d5d-bec3-4e00-81ed-4fbd57215d78 req-e2bd92ac-22b5-4b11-b3be-9680f8d5ca08 service nova] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Received event network-vif-deleted-088f6f2e-acc0-4966-9ee5-47b6c6799c5b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.943945] env[63197]: DEBUG nova.compute.manager [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Received event network-vif-deleted-5685fd2b-99cb-4b1e-b61d-9ff2f290f33d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.944220] env[63197]: DEBUG nova.compute.manager [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-vif-plugged-4403e435-3845-48ec-a898-a1a932fea8d3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.944447] env[63197]: DEBUG oslo_concurrency.lockutils [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] Acquiring lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.944680] env[63197]: DEBUG oslo_concurrency.lockutils [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.944877] env[63197]: DEBUG oslo_concurrency.lockutils [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.945077] env[63197]: DEBUG nova.compute.manager [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] No waiting events found dispatching network-vif-plugged-4403e435-3845-48ec-a898-a1a932fea8d3 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 909.945277] env[63197]: WARNING nova.compute.manager [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received unexpected event network-vif-plugged-4403e435-3845-48ec-a898-a1a932fea8d3 for instance with vm_state building and task_state spawning. [ 909.945464] env[63197]: DEBUG nova.compute.manager [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-changed-4403e435-3845-48ec-a898-a1a932fea8d3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.945645] env[63197]: DEBUG nova.compute.manager [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Refreshing instance network info cache due to event network-changed-4403e435-3845-48ec-a898-a1a932fea8d3. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.945836] env[63197]: DEBUG oslo_concurrency.lockutils [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] Acquiring lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.945972] env[63197]: DEBUG oslo_concurrency.lockutils [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] Acquired lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.946143] env[63197]: DEBUG nova.network.neutron [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Refreshing network info cache for port 4403e435-3845-48ec-a898-a1a932fea8d3 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.968805] env[63197]: DEBUG nova.compute.manager [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.969812] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d51654-08be-4f24-a80a-1c9e4c0e6c20 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.313384] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4321f027-3323-49ba-b362-33ffa178d8eb tempest-MultipleCreateTestJSON-2002895911 tempest-MultipleCreateTestJSON-2002895911-project-member] Lock "8500a7e9-3fdc-411e-a48e-189d4d7bffba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.225s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.375235] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52203c96-aa0d-cce4-e0a4-4118cfad2437, 'name': SearchDatastore_Task, 'duration_secs': 0.04052} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.375522] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.375788] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] a1f365de-b634-481a-b2e8-9bfc801a47f5/a1f365de-b634-481a-b2e8-9bfc801a47f5.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.377623] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-542c3525-1e45-4c75-b040-f4a76eb2fcab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.383567] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 910.383567] env[63197]: value = "task-1364262" [ 910.383567] env[63197]: _type = "Task" [ 910.383567] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.392521] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.484168] env[63197]: INFO nova.compute.manager [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] instance snapshotting [ 910.486323] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbe9442-1781-48e2-b5bb-b421ebcf8709 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.492186] env[63197]: DEBUG nova.compute.manager [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.492186] env[63197]: DEBUG nova.compute.manager [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing instance network info cache due to event network-changed-5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 910.492186] env[63197]: DEBUG oslo_concurrency.lockutils [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] Acquiring lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.492336] env[63197]: DEBUG oslo_concurrency.lockutils [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] Acquired lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.492417] env[63197]: DEBUG nova.network.neutron [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Refreshing network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 910.511472] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed26c6b-73ea-41e1-b279-3d23795b007e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.524122] env[63197]: DEBUG nova.network.neutron [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.624057] env[63197]: DEBUG nova.network.neutron [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.897627] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364262, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.001785] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8d5844-d237-4429-84b9-d8308d398757 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.009220] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf2f306-c43e-4a7e-92b7-2eb214a6b712 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.042332] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 911.043234] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4b096da4-5782-497d-a2ef-59f93ccdb994 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.045848] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba316ef4-9483-4181-b026-4dd30cf1c63a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.055688] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385a6c02-a2e5-4aaa-b8d0-acb06b637227 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.060354] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 911.060354] env[63197]: value = "task-1364263" [ 911.060354] env[63197]: _type = "Task" [ 911.060354] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.071365] env[63197]: DEBUG nova.compute.provider_tree [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.077943] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364263, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.126774] env[63197]: DEBUG oslo_concurrency.lockutils [req-6b33730f-af92-42f5-aae2-b6d3acac98fc req-1928eebc-92ad-40af-9df8-d07b8fe09b20 service nova] Releasing lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.291597] env[63197]: DEBUG nova.network.neutron [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updated VIF entry in instance network info cache for port 5f405bfb-7d3e-4876-8fe0-2dee955c3325. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 911.292057] env[63197]: DEBUG nova.network.neutron [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [{"id": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "address": "fa:16:3e:c5:df:50", "network": {"id": "564d5eda-ffb0-4fbf-986f-ea1c373b362e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1269026000-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558e5440e64c43f59d0f2cc54c18c79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee4b2432-c393-4e50-ae0e-b5e12bad37db", "external-id": "nsx-vlan-transportzone-985", "segmentation_id": 985, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f405bfb-7d", "ovs_interfaceid": "5f405bfb-7d3e-4876-8fe0-2dee955c3325", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.396542] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547474} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.396846] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] a1f365de-b634-481a-b2e8-9bfc801a47f5/a1f365de-b634-481a-b2e8-9bfc801a47f5.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.397017] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.397276] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e689048a-3cd2-4b7f-a1e2-264ba5b245ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.403763] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 911.403763] env[63197]: value = "task-1364264" [ 911.403763] env[63197]: _type = "Task" [ 911.403763] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.412591] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364264, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.578864] env[63197]: DEBUG nova.scheduler.client.report [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.583370] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364263, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.795820] env[63197]: DEBUG oslo_concurrency.lockutils [req-11299736-253f-4c15-b865-fe0b406b7482 req-6a90ac26-c255-4221-a9fd-3cc637de9e42 service nova] Releasing lock "refresh_cache-65df32fa-a6e7-4c3a-af8b-0422df8d9229" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.886721] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Successfully updated port: cf42dede-8506-413b-a119-cbe31f3be836 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.914471] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364264, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067994} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.914983] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.915764] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce18b0a-a14d-4b25-a71b-9c6bc9386711 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.936948] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] a1f365de-b634-481a-b2e8-9bfc801a47f5/a1f365de-b634-481a-b2e8-9bfc801a47f5.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.937478] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b31d3f8-7393-405a-8ee7-0afb130c6f06 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.956723] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 911.956723] env[63197]: value = "task-1364265" [ 911.956723] env[63197]: _type = "Task" [ 911.956723] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.964116] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.969794] env[63197]: DEBUG nova.compute.manager [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-vif-plugged-cf42dede-8506-413b-a119-cbe31f3be836 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.969997] env[63197]: DEBUG oslo_concurrency.lockutils [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] Acquiring lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.970223] env[63197]: DEBUG oslo_concurrency.lockutils [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.970396] env[63197]: DEBUG oslo_concurrency.lockutils [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.970556] env[63197]: DEBUG nova.compute.manager [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] No waiting events found dispatching network-vif-plugged-cf42dede-8506-413b-a119-cbe31f3be836 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.970712] env[63197]: WARNING nova.compute.manager [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received unexpected event network-vif-plugged-cf42dede-8506-413b-a119-cbe31f3be836 for instance with vm_state building and task_state spawning. [ 911.970931] env[63197]: DEBUG nova.compute.manager [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-changed-cf42dede-8506-413b-a119-cbe31f3be836 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.971147] env[63197]: DEBUG nova.compute.manager [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Refreshing instance network info cache due to event network-changed-cf42dede-8506-413b-a119-cbe31f3be836. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.971376] env[63197]: DEBUG oslo_concurrency.lockutils [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] Acquiring lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.971533] env[63197]: DEBUG oslo_concurrency.lockutils [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] Acquired lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.971712] env[63197]: DEBUG nova.network.neutron [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Refreshing network info cache for port cf42dede-8506-413b-a119-cbe31f3be836 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 912.071480] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364263, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.085540] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.313s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.086083] env[63197]: DEBUG nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.088651] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.015s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.088845] env[63197]: DEBUG nova.objects.instance [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63197) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 912.350391] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4c31dc4-19bc-4ec0-8ff9-413b6144d98f tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.350699] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4c31dc4-19bc-4ec0-8ff9-413b6144d98f tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.351031] env[63197]: DEBUG nova.objects.instance [None req-f4c31dc4-19bc-4ec0-8ff9-413b6144d98f tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'flavor' on Instance uuid 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.467313] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364265, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.547762] env[63197]: DEBUG nova.network.neutron [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.575022] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364263, 'name': CreateSnapshot_Task, 'duration_secs': 1.139458} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.575022] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 912.575022] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2a7831-386a-47c1-a641-06b56bfa7577 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.595576] env[63197]: DEBUG nova.compute.utils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.595576] env[63197]: DEBUG nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 912.595576] env[63197]: DEBUG nova.network.neutron [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 912.655693] env[63197]: DEBUG nova.policy [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81fc34f0c73c4baf829c5f9cae7510fc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d992f8d3d334b8783f3dc5eff1cf423', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.691674] env[63197]: DEBUG nova.network.neutron [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.854527] env[63197]: DEBUG nova.objects.instance [None req-f4c31dc4-19bc-4ec0-8ff9-413b6144d98f tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'pci_requests' on Instance uuid 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.906423] env[63197]: DEBUG nova.network.neutron [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Successfully created port: dd634e1d-f13f-4709-a49b-4a669d1286f3 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.968047] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364265, 'name': ReconfigVM_Task, 'duration_secs': 0.80779} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.968345] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Reconfigured VM instance instance-0000004b to attach disk [datastore1] a1f365de-b634-481a-b2e8-9bfc801a47f5/a1f365de-b634-481a-b2e8-9bfc801a47f5.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.969120] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a41b1e05-e7ac-4f43-a448-8988b2819790 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.975855] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 912.975855] env[63197]: value = "task-1364266" [ 912.975855] env[63197]: _type = "Task" [ 912.975855] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.983884] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364266, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.095980] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 913.096347] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-af846d21-2928-4df3-a57a-2a7916d63a17 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.105331] env[63197]: DEBUG nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.110123] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 913.110123] env[63197]: value = "task-1364267" [ 913.110123] env[63197]: _type = "Task" [ 913.110123] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.110123] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3c30afe2-c9d8-4a2d-9046-e4d617da3da2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.110866] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.204s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.111892] env[63197]: INFO nova.compute.claims [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.124644] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364267, 'name': CloneVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.193561] env[63197]: DEBUG oslo_concurrency.lockutils [req-6da2f266-2cef-4eb5-a405-dd09a7eb59d2 req-531917b4-a31d-4250-ba9e-c475d4af5a4a service nova] Releasing lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.357637] env[63197]: DEBUG nova.objects.base [None req-f4c31dc4-19bc-4ec0-8ff9-413b6144d98f tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Object Instance<9c1caa2b-c369-425a-8726-cddadf06f338> lazy-loaded attributes: flavor,pci_requests {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 913.357853] env[63197]: DEBUG nova.network.neutron [None req-f4c31dc4-19bc-4ec0-8ff9-413b6144d98f tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.476867] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f4c31dc4-19bc-4ec0-8ff9-413b6144d98f tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.126s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.487535] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364266, 'name': Rename_Task, 'duration_secs': 0.172814} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.488432] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.488688] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a07487c-d1a6-4af5-8632-3e4eae2cead1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.496049] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 913.496049] env[63197]: value = "task-1364268" [ 913.496049] env[63197]: _type = "Task" [ 913.496049] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.504067] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.623883] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364267, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.010378] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364268, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.064371] env[63197]: DEBUG nova.compute.manager [req-d9adefc0-74f7-4f31-8ee0-9c9378a7c175 req-c4c95342-e5cb-4c7e-9922-c491a06fff50 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-vif-plugged-c856ee9c-0a79-46b8-a8ec-f6283349a05e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.064371] env[63197]: DEBUG oslo_concurrency.lockutils [req-d9adefc0-74f7-4f31-8ee0-9c9378a7c175 req-c4c95342-e5cb-4c7e-9922-c491a06fff50 service nova] Acquiring lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.064371] env[63197]: DEBUG oslo_concurrency.lockutils [req-d9adefc0-74f7-4f31-8ee0-9c9378a7c175 req-c4c95342-e5cb-4c7e-9922-c491a06fff50 service nova] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.064371] env[63197]: DEBUG oslo_concurrency.lockutils [req-d9adefc0-74f7-4f31-8ee0-9c9378a7c175 req-c4c95342-e5cb-4c7e-9922-c491a06fff50 service nova] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.064371] env[63197]: DEBUG nova.compute.manager [req-d9adefc0-74f7-4f31-8ee0-9c9378a7c175 req-c4c95342-e5cb-4c7e-9922-c491a06fff50 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] No waiting events found dispatching network-vif-plugged-c856ee9c-0a79-46b8-a8ec-f6283349a05e {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.064534] env[63197]: WARNING nova.compute.manager [req-d9adefc0-74f7-4f31-8ee0-9c9378a7c175 req-c4c95342-e5cb-4c7e-9922-c491a06fff50 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received unexpected event network-vif-plugged-c856ee9c-0a79-46b8-a8ec-f6283349a05e for instance with vm_state building and task_state spawning. [ 914.119454] env[63197]: DEBUG nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.125044] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Successfully updated port: c856ee9c-0a79-46b8-a8ec-f6283349a05e {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.135132] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364267, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.151324] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.151651] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.151909] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.152243] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.152480] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.152727] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.153092] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.153341] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.153596] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.153829] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.154114] env[63197]: DEBUG nova.virt.hardware [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.155236] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d307ad48-7fea-434d-b0c0-9f7f62e37ed2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.166530] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51a8949-52a2-4329-a12b-b0fa40018c1f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.304256] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b043190a-7187-4c4b-bf22-fcce7c626013 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.311928] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea166aa-b958-4afd-bb88-a659ff84de23 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.340746] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-543232a9-ae21-47b5-a8e5-989f75f803e0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.347767] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce104eda-3404-4e84-bac3-ed8884a8e57d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.362030] env[63197]: DEBUG nova.compute.provider_tree [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.505451] env[63197]: DEBUG oslo_vmware.api [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364268, 'name': PowerOnVM_Task, 'duration_secs': 0.528131} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.505896] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.505896] env[63197]: INFO nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Took 10.69 seconds to spawn the instance on the hypervisor. [ 914.506129] env[63197]: DEBUG nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 914.506965] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9a6e62-1e54-4aa7-a171-bab008b508d6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.587768] env[63197]: DEBUG nova.compute.manager [req-0d8c859d-fc3b-4ba4-96a9-e2a5e5ff271d req-a83c8f22-0595-4f20-b134-31bed7f8a7c0 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Received event network-vif-plugged-dd634e1d-f13f-4709-a49b-4a669d1286f3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.587978] env[63197]: DEBUG oslo_concurrency.lockutils [req-0d8c859d-fc3b-4ba4-96a9-e2a5e5ff271d req-a83c8f22-0595-4f20-b134-31bed7f8a7c0 service nova] Acquiring lock "52549788-b58f-4f5d-8443-a0c771099884-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.588202] env[63197]: DEBUG oslo_concurrency.lockutils [req-0d8c859d-fc3b-4ba4-96a9-e2a5e5ff271d req-a83c8f22-0595-4f20-b134-31bed7f8a7c0 service nova] Lock "52549788-b58f-4f5d-8443-a0c771099884-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.588404] env[63197]: DEBUG oslo_concurrency.lockutils [req-0d8c859d-fc3b-4ba4-96a9-e2a5e5ff271d req-a83c8f22-0595-4f20-b134-31bed7f8a7c0 service nova] Lock "52549788-b58f-4f5d-8443-a0c771099884-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.588741] env[63197]: DEBUG nova.compute.manager [req-0d8c859d-fc3b-4ba4-96a9-e2a5e5ff271d req-a83c8f22-0595-4f20-b134-31bed7f8a7c0 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] No waiting events found dispatching network-vif-plugged-dd634e1d-f13f-4709-a49b-4a669d1286f3 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.588915] env[63197]: WARNING nova.compute.manager [req-0d8c859d-fc3b-4ba4-96a9-e2a5e5ff271d req-a83c8f22-0595-4f20-b134-31bed7f8a7c0 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Received unexpected event network-vif-plugged-dd634e1d-f13f-4709-a49b-4a669d1286f3 for instance with vm_state building and task_state spawning. [ 914.624262] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364267, 'name': CloneVM_Task} progress is 95%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.625159] env[63197]: DEBUG nova.network.neutron [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Successfully updated port: dd634e1d-f13f-4709-a49b-4a669d1286f3 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.631930] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.632101] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.632256] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.866284] env[63197]: DEBUG nova.scheduler.client.report [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.025144] env[63197]: INFO nova.compute.manager [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Took 29.98 seconds to build instance. [ 915.124829] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364267, 'name': CloneVM_Task, 'duration_secs': 1.69269} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.125060] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Created linked-clone VM from snapshot [ 915.125818] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ce6c92-be91-40cd-9574-f57548cc0ad6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.129117] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.129288] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.129469] env[63197]: DEBUG nova.network.neutron [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.137365] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Uploading image 317ea0a5-01ce-4617-891d-f3a00c993dd1 {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 915.149969] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 915.150269] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d186e865-1163-4f72-85f8-d6453ebfa5b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.157603] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 915.157603] env[63197]: value = "task-1364269" [ 915.157603] env[63197]: _type = "Task" [ 915.157603] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.165556] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364269, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.183740] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 915.371061] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.371586] env[63197]: DEBUG nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.376294] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.108s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.376481] env[63197]: DEBUG nova.objects.instance [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lazy-loading 'resources' on Instance uuid 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.427965] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.428240] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.428554] env[63197]: DEBUG nova.objects.instance [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'flavor' on Instance uuid 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.527387] env[63197]: DEBUG oslo_concurrency.lockutils [None req-74563bd0-5caf-42f7-8052-43d3f10ae67f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.788s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.668060] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364269, 'name': Destroy_Task, 'duration_secs': 0.285244} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.669238] env[63197]: DEBUG nova.network.neutron [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 915.671067] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Destroyed the VM [ 915.671699] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 915.672008] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0baac7ac-ff79-41b2-8606-15344dfa770d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.678867] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 915.678867] env[63197]: value = "task-1364270" [ 915.678867] env[63197]: _type = "Task" [ 915.678867] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.686631] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364270, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.720186] env[63197]: DEBUG nova.network.neutron [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Updating instance_info_cache with network_info: [{"id": "4403e435-3845-48ec-a898-a1a932fea8d3", "address": "fa:16:3e:7c:87:64", "network": {"id": "f63fd958-33f7-4275-ad22-bb6545390ccc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1942842288", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.38", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4403e435-38", "ovs_interfaceid": "4403e435-3845-48ec-a898-a1a932fea8d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf42dede-8506-413b-a119-cbe31f3be836", "address": "fa:16:3e:2e:81:9f", "network": {"id": "1fe2f14b-d6c0-42f3-a4e1-f70a3cf56e6d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1080454701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42dede-85", "ovs_interfaceid": "cf42dede-8506-413b-a119-cbe31f3be836", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "address": "fa:16:3e:cb:2b:5e", "network": {"id": "f63fd958-33f7-4275-ad22-bb6545390ccc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1942842288", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc856ee9c-0a", "ovs_interfaceid": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.799851] env[63197]: DEBUG nova.network.neutron [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Updating instance_info_cache with network_info: [{"id": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "address": "fa:16:3e:b8:5c:89", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd634e1d-f1", "ovs_interfaceid": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.879904] env[63197]: DEBUG nova.compute.utils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.885858] env[63197]: DEBUG nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.885858] env[63197]: DEBUG nova.network.neutron [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 915.938456] env[63197]: DEBUG nova.policy [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b50be4b75a94b4481c9c65ea1e4e9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bb89fb32d8c4726a9a3104d68ce560a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.033626] env[63197]: DEBUG nova.objects.instance [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'pci_requests' on Instance uuid 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.074134] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551d9034-5475-4fd2-a35d-cfb75cc49ed0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.082045] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a08320c-2056-4552-b7ea-6c15ce0c157f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.114615] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f439c35b-5aa2-401e-accb-ec19c5480e94 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.121793] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb81633-e501-44b5-895f-6fc4d98c5fad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.134762] env[63197]: DEBUG nova.compute.provider_tree [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.190636] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364270, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.210883] env[63197]: DEBUG nova.network.neutron [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Successfully created port: 2d53dd8b-1ba1-495d-99fc-265d17cb34a5 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.222875] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Releasing lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.223413] env[63197]: DEBUG nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Instance network_info: |[{"id": "4403e435-3845-48ec-a898-a1a932fea8d3", "address": "fa:16:3e:7c:87:64", "network": {"id": "f63fd958-33f7-4275-ad22-bb6545390ccc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1942842288", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.38", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4403e435-38", "ovs_interfaceid": "4403e435-3845-48ec-a898-a1a932fea8d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf42dede-8506-413b-a119-cbe31f3be836", "address": "fa:16:3e:2e:81:9f", "network": {"id": "1fe2f14b-d6c0-42f3-a4e1-f70a3cf56e6d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1080454701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42dede-85", "ovs_interfaceid": "cf42dede-8506-413b-a119-cbe31f3be836", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "address": "fa:16:3e:cb:2b:5e", "network": {"id": "f63fd958-33f7-4275-ad22-bb6545390ccc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1942842288", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc856ee9c-0a", "ovs_interfaceid": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 916.223830] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:87:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ba07329-1d3e-4ba8-8774-d029262318c4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4403e435-3845-48ec-a898-a1a932fea8d3', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:81:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c1520c99-af74-4d61-a8ae-56aef56ef4f0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf42dede-8506-413b-a119-cbe31f3be836', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:2b:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ba07329-1d3e-4ba8-8774-d029262318c4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c856ee9c-0a79-46b8-a8ec-f6283349a05e', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.239018] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Creating folder: Project (e4f082d2f82c42218116efa35aee1f56). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 916.239018] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9314bd36-bf4a-41c0-a9ce-5646429c6480 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.246711] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Created folder: Project (e4f082d2f82c42218116efa35aee1f56) in parent group-v290286. [ 916.246890] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Creating folder: Instances. Parent ref: group-v290354. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 916.247144] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f07f46d0-6e46-4457-be0d-ad61c8f140eb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.255903] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Created folder: Instances in parent group-v290354. [ 916.256198] env[63197]: DEBUG oslo.service.loopingcall [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.256405] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 916.256626] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0d980c9f-7ecc-4533-861c-c8e073a6733c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.280384] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.280384] env[63197]: value = "task-1364273" [ 916.280384] env[63197]: _type = "Task" [ 916.280384] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.289604] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364273, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.303290] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.303614] env[63197]: DEBUG nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Instance network_info: |[{"id": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "address": "fa:16:3e:b8:5c:89", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd634e1d-f1", "ovs_interfaceid": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 916.304032] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:5c:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089ef678-58b4-4bf0-a39d-b94b2d364291', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd634e1d-f13f-4709-a49b-4a669d1286f3', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.311360] env[63197]: DEBUG oslo.service.loopingcall [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.311609] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 916.311859] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e08845b8-0296-420c-bc3a-6b7c71791892 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.330289] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.330289] env[63197]: value = "task-1364274" [ 916.330289] env[63197]: _type = "Task" [ 916.330289] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.335086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "a1f365de-b634-481a-b2e8-9bfc801a47f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.335323] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.335524] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "a1f365de-b634-481a-b2e8-9bfc801a47f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.335700] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.335864] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.340941] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364274, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.341653] env[63197]: INFO nova.compute.manager [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Terminating instance [ 916.344906] env[63197]: DEBUG nova.compute.manager [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-changed-c856ee9c-0a79-46b8-a8ec-f6283349a05e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.345107] env[63197]: DEBUG nova.compute.manager [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Refreshing instance network info cache due to event network-changed-c856ee9c-0a79-46b8-a8ec-f6283349a05e. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.345323] env[63197]: DEBUG oslo_concurrency.lockutils [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] Acquiring lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.345500] env[63197]: DEBUG oslo_concurrency.lockutils [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] Acquired lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.345784] env[63197]: DEBUG nova.network.neutron [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Refreshing network info cache for port c856ee9c-0a79-46b8-a8ec-f6283349a05e {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 916.347224] env[63197]: DEBUG nova.compute.manager [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.347412] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.348442] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee32fd0-fc0c-4063-9d79-7f179df5be78 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.355869] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.356340] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f46f2ee9-5309-40f3-ade3-b9fd7966931a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.362741] env[63197]: DEBUG oslo_vmware.api [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 916.362741] env[63197]: value = "task-1364275" [ 916.362741] env[63197]: _type = "Task" [ 916.362741] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.370555] env[63197]: DEBUG oslo_vmware.api [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.386276] env[63197]: DEBUG nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.536607] env[63197]: DEBUG nova.objects.base [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Object Instance<9c1caa2b-c369-425a-8726-cddadf06f338> lazy-loaded attributes: flavor,pci_requests {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 916.537069] env[63197]: DEBUG nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 916.577539] env[63197]: DEBUG nova.policy [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.616222] env[63197]: DEBUG nova.compute.manager [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Received event network-changed-dd634e1d-f13f-4709-a49b-4a669d1286f3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.616368] env[63197]: DEBUG nova.compute.manager [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Refreshing instance network info cache due to event network-changed-dd634e1d-f13f-4709-a49b-4a669d1286f3. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.616585] env[63197]: DEBUG oslo_concurrency.lockutils [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] Acquiring lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.616725] env[63197]: DEBUG oslo_concurrency.lockutils [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] Acquired lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.616890] env[63197]: DEBUG nova.network.neutron [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Refreshing network info cache for port dd634e1d-f13f-4709-a49b-4a669d1286f3 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 916.637642] env[63197]: DEBUG nova.scheduler.client.report [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.690226] env[63197]: DEBUG oslo_vmware.api [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364270, 'name': RemoveSnapshot_Task, 'duration_secs': 0.637152} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.690421] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 916.791742] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364273, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.839424] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364274, 'name': CreateVM_Task, 'duration_secs': 0.395629} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.839594] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.840318] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.840472] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.840782] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.841044] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-287ccb33-bc87-4dad-845c-0954a6695539 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.845343] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 916.845343] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c7d6ef-2059-1b8d-0663-b03f577d2f85" [ 916.845343] env[63197]: _type = "Task" [ 916.845343] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.854624] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c7d6ef-2059-1b8d-0663-b03f577d2f85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.875755] env[63197]: DEBUG oslo_vmware.api [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364275, 'name': PowerOffVM_Task, 'duration_secs': 0.194175} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.875755] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.875755] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 916.875755] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1db47946-79d3-4e35-9a53-889db2244e72 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.897340] env[63197]: DEBUG nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Successfully created port: 9dfe2781-a394-4232-ac80-7e536c691f94 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.942455] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 916.942671] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 916.942850] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleting the datastore file [datastore1] a1f365de-b634-481a-b2e8-9bfc801a47f5 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.943137] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7362375-5212-467c-9a7c-90d821a56256 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.949780] env[63197]: DEBUG oslo_vmware.api [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 916.949780] env[63197]: value = "task-1364277" [ 916.949780] env[63197]: _type = "Task" [ 916.949780] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.960038] env[63197]: DEBUG oslo_vmware.api [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364277, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.076982] env[63197]: DEBUG nova.network.neutron [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Updated VIF entry in instance network info cache for port c856ee9c-0a79-46b8-a8ec-f6283349a05e. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 917.077128] env[63197]: DEBUG nova.network.neutron [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Updating instance_info_cache with network_info: [{"id": "4403e435-3845-48ec-a898-a1a932fea8d3", "address": "fa:16:3e:7c:87:64", "network": {"id": "f63fd958-33f7-4275-ad22-bb6545390ccc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1942842288", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.38", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4403e435-38", "ovs_interfaceid": "4403e435-3845-48ec-a898-a1a932fea8d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cf42dede-8506-413b-a119-cbe31f3be836", "address": "fa:16:3e:2e:81:9f", "network": {"id": "1fe2f14b-d6c0-42f3-a4e1-f70a3cf56e6d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1080454701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42dede-85", "ovs_interfaceid": "cf42dede-8506-413b-a119-cbe31f3be836", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "address": "fa:16:3e:cb:2b:5e", "network": {"id": "f63fd958-33f7-4275-ad22-bb6545390ccc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1942842288", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc856ee9c-0a", "ovs_interfaceid": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.144170] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.147106] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.693s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.147351] env[63197]: DEBUG nova.objects.instance [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lazy-loading 'resources' on Instance uuid 232ea430-6a11-4dbc-9da9-0f734471d03b {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.167330] env[63197]: INFO nova.scheduler.client.report [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted allocations for instance 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d [ 917.195246] env[63197]: WARNING nova.compute.manager [None req-70dea080-99a9-40a0-ac01-ed16a57ea9d1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Image not found during snapshot: nova.exception.ImageNotFound: Image 317ea0a5-01ce-4617-891d-f3a00c993dd1 could not be found. [ 917.292629] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364273, 'name': CreateVM_Task, 'duration_secs': 0.549019} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.292801] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 917.293705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.319658] env[63197]: DEBUG nova.network.neutron [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Updated VIF entry in instance network info cache for port dd634e1d-f13f-4709-a49b-4a669d1286f3. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 917.320102] env[63197]: DEBUG nova.network.neutron [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Updating instance_info_cache with network_info: [{"id": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "address": "fa:16:3e:b8:5c:89", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd634e1d-f1", "ovs_interfaceid": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.357868] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c7d6ef-2059-1b8d-0663-b03f577d2f85, 'name': SearchDatastore_Task, 'duration_secs': 0.008719} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.358197] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.358420] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.358645] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.358784] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.358953] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.359291] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.359598] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.359818] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-698c472b-3a8a-4390-8425-d2456bd533ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.361632] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5737164-a1a6-4b3c-9b71-515e8fb5b8df {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.367425] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 917.367425] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5203f2a3-e147-37fa-cdf5-48188e553bc7" [ 917.367425] env[63197]: _type = "Task" [ 917.367425] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.371394] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.371562] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.372562] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-58a57641-dd1b-41d3-8da1-17b730e060d9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.377447] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5203f2a3-e147-37fa-cdf5-48188e553bc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.380203] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 917.380203] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527d8e7f-b600-9fc2-5d31-0f6a3b32d4cb" [ 917.380203] env[63197]: _type = "Task" [ 917.380203] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.387640] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527d8e7f-b600-9fc2-5d31-0f6a3b32d4cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.396831] env[63197]: DEBUG nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.422434] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.422674] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.422826] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.423009] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.423201] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.423366] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.423570] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.423723] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.423882] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.424055] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.424236] env[63197]: DEBUG nova.virt.hardware [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.425158] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8cce9b-05bb-4cc7-9ba9-92560edac336 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.433210] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214acc99-1789-49ba-ae47-bf606ed4f464 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.458316] env[63197]: DEBUG oslo_vmware.api [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364277, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.364211} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.458552] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.458731] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.458902] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.459108] env[63197]: INFO nova.compute.manager [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 917.459349] env[63197]: DEBUG oslo.service.loopingcall [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.459538] env[63197]: DEBUG nova.compute.manager [-] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 917.459631] env[63197]: DEBUG nova.network.neutron [-] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.579475] env[63197]: DEBUG oslo_concurrency.lockutils [req-0751a0ea-cb6d-4a88-b68d-4e1dc7f2f43a req-0d6c1a1d-f4a8-4945-b341-c9dfadae7eda service nova] Releasing lock "refresh_cache-e87c0992-0c7c-4de8-ac68-800eb80cfbc4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.677848] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3eb17ec2-76f6-43dc-bfc9-ad7229c31969 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "1067806f-7a4d-4e3f-86c5-7b0d97f33d2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.588s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.825908] env[63197]: DEBUG oslo_concurrency.lockutils [req-8bd34458-a1fd-4368-a6e2-891100c45ce4 req-090040c8-2a1b-4f33-8199-f1f2e0c935b9 service nova] Releasing lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.861357] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4237206a-74bf-4b3e-b899-fda990e210ce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.879120] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70c723a-d824-43ae-a75d-e2b96357e2e2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.890184] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5203f2a3-e147-37fa-cdf5-48188e553bc7, 'name': SearchDatastore_Task, 'duration_secs': 0.008303} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.895164] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.895538] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.895573] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.896828] env[63197]: DEBUG nova.network.neutron [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Successfully updated port: 2d53dd8b-1ba1-495d-99fc-265d17cb34a5 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 917.931515] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8449d145-eef8-42fc-b017-a2de2bc955ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.938251] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527d8e7f-b600-9fc2-5d31-0f6a3b32d4cb, 'name': SearchDatastore_Task, 'duration_secs': 0.007524} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.939529] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ce2be14-a5e9-485c-8c23-4284f0926e1c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.946937] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5db1d86-0c89-4446-8910-56f8c996033f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.953181] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 917.953181] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52429ab0-6fcd-c664-421d-5e04b683f7ca" [ 917.953181] env[63197]: _type = "Task" [ 917.953181] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.966347] env[63197]: DEBUG nova.compute.provider_tree [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.978027] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52429ab0-6fcd-c664-421d-5e04b683f7ca, 'name': SearchDatastore_Task, 'duration_secs': 0.022836} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.978027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.978027] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/52549788-b58f-4f5d-8443-a0c771099884.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.978027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.978027] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.978027] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0949792-0b2e-42fe-8c7b-c5333823b247 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.978868] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d31ee2c6-da33-4fdd-a444-c7bfdb5b5a06 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.987557] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 917.987557] env[63197]: value = "task-1364278" [ 917.987557] env[63197]: _type = "Task" [ 917.987557] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.988831] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.989027] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.993067] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81874f3a-6fb1-46dc-8457-e1ac10c304fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.002348] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364278, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.002659] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 918.002659] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52374a4f-1dab-86a3-7e47-78bad90a27fc" [ 918.002659] env[63197]: _type = "Task" [ 918.002659] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.011385] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52374a4f-1dab-86a3-7e47-78bad90a27fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.103590] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "9d8c131d-1183-4508-ae2d-28e38a50e58d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.103905] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.104137] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "9d8c131d-1183-4508-ae2d-28e38a50e58d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.104326] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.104496] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.106706] env[63197]: INFO nova.compute.manager [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Terminating instance [ 918.108446] env[63197]: DEBUG nova.compute.manager [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 918.108634] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 918.109453] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836b3390-c46a-4ceb-ade5-c742b76e839b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.117909] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 918.118186] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f3ac035e-5628-4275-bfd0-96ae3a0d47bf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.125932] env[63197]: DEBUG oslo_vmware.api [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 918.125932] env[63197]: value = "task-1364279" [ 918.125932] env[63197]: _type = "Task" [ 918.125932] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.134667] env[63197]: DEBUG oslo_vmware.api [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.350466] env[63197]: DEBUG nova.network.neutron [-] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.401523] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.401869] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.402222] env[63197]: DEBUG nova.network.neutron [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.470171] env[63197]: DEBUG nova.scheduler.client.report [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.479785] env[63197]: DEBUG nova.compute.manager [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Received event network-vif-plugged-2d53dd8b-1ba1-495d-99fc-265d17cb34a5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.479958] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] Acquiring lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.480153] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.480365] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.480721] env[63197]: DEBUG nova.compute.manager [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] No waiting events found dispatching network-vif-plugged-2d53dd8b-1ba1-495d-99fc-265d17cb34a5 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.480824] env[63197]: WARNING nova.compute.manager [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Received unexpected event network-vif-plugged-2d53dd8b-1ba1-495d-99fc-265d17cb34a5 for instance with vm_state building and task_state spawning. [ 918.480961] env[63197]: DEBUG nova.compute.manager [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Received event network-vif-deleted-5a2fa3ff-121f-40d3-b800-77904777a6d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.482099] env[63197]: DEBUG nova.compute.manager [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Received event network-changed-2d53dd8b-1ba1-495d-99fc-265d17cb34a5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.482099] env[63197]: DEBUG nova.compute.manager [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Refreshing instance network info cache due to event network-changed-2d53dd8b-1ba1-495d-99fc-265d17cb34a5. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.482099] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] Acquiring lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.500846] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364278, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.521287] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52374a4f-1dab-86a3-7e47-78bad90a27fc, 'name': SearchDatastore_Task, 'duration_secs': 0.012738} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.522182] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-350a9103-c101-4931-b3d3-99a7652b9f05 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.533146] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 918.533146] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529b41d7-a7b9-ed24-94bb-88133f5011c5" [ 918.533146] env[63197]: _type = "Task" [ 918.533146] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.542205] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529b41d7-a7b9-ed24-94bb-88133f5011c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.570168] env[63197]: DEBUG nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Successfully updated port: 9dfe2781-a394-4232-ac80-7e536c691f94 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.638995] env[63197]: DEBUG oslo_vmware.api [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364279, 'name': PowerOffVM_Task, 'duration_secs': 0.247592} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.639488] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.639685] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.640131] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd07497a-f97c-428a-adf4-ab84e1c7cef5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.646622] env[63197]: DEBUG nova.compute.manager [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-vif-plugged-9dfe2781-a394-4232-ac80-7e536c691f94 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.646931] env[63197]: DEBUG oslo_concurrency.lockutils [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.647277] env[63197]: DEBUG oslo_concurrency.lockutils [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.647530] env[63197]: DEBUG oslo_concurrency.lockutils [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.647771] env[63197]: DEBUG nova.compute.manager [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] No waiting events found dispatching network-vif-plugged-9dfe2781-a394-4232-ac80-7e536c691f94 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.648034] env[63197]: WARNING nova.compute.manager [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received unexpected event network-vif-plugged-9dfe2781-a394-4232-ac80-7e536c691f94 for instance with vm_state active and task_state None. [ 918.648283] env[63197]: DEBUG nova.compute.manager [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-changed-9dfe2781-a394-4232-ac80-7e536c691f94 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.648513] env[63197]: DEBUG nova.compute.manager [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing instance network info cache due to event network-changed-9dfe2781-a394-4232-ac80-7e536c691f94. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 918.648783] env[63197]: DEBUG oslo_concurrency.lockutils [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.648989] env[63197]: DEBUG oslo_concurrency.lockutils [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.649236] env[63197]: DEBUG nova.network.neutron [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing network info cache for port 9dfe2781-a394-4232-ac80-7e536c691f94 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.729698] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.729971] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.730193] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Deleting the datastore file [datastore1] 9d8c131d-1183-4508-ae2d-28e38a50e58d {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.730485] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f2baa10-d9a7-473a-8d52-623b7cb9f49c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.737451] env[63197]: DEBUG oslo_vmware.api [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for the task: (returnval){ [ 918.737451] env[63197]: value = "task-1364281" [ 918.737451] env[63197]: _type = "Task" [ 918.737451] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.745762] env[63197]: DEBUG oslo_vmware.api [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364281, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.854097] env[63197]: INFO nova.compute.manager [-] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Took 1.39 seconds to deallocate network for instance. [ 918.936370] env[63197]: DEBUG nova.network.neutron [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.979567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.979567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.131s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.001395] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364278, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699825} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.001395] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/52549788-b58f-4f5d-8443-a0c771099884.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 919.001866] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.001866] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0bafdbcb-bd6a-4ff4-a018-6d9617684b67 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.008545] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 919.008545] env[63197]: value = "task-1364282" [ 919.008545] env[63197]: _type = "Task" [ 919.008545] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.017828] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364282, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.019282] env[63197]: INFO nova.scheduler.client.report [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Deleted allocations for instance 232ea430-6a11-4dbc-9da9-0f734471d03b [ 919.045403] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529b41d7-a7b9-ed24-94bb-88133f5011c5, 'name': SearchDatastore_Task, 'duration_secs': 0.05846} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.047014] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.047321] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] e87c0992-0c7c-4de8-ac68-800eb80cfbc4/e87c0992-0c7c-4de8-ac68-800eb80cfbc4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.047777] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8125a76a-234c-4236-9ad5-4900e5a36568 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.055311] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 919.055311] env[63197]: value = "task-1364283" [ 919.055311] env[63197]: _type = "Task" [ 919.055311] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.064427] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364283, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.081691] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.126600] env[63197]: DEBUG nova.network.neutron [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance_info_cache with network_info: [{"id": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "address": "fa:16:3e:be:d6:04", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d53dd8b-1b", "ovs_interfaceid": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.252926] env[63197]: DEBUG oslo_vmware.api [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Task: {'id': task-1364281, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135795} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.253265] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.253460] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 919.253638] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 919.253807] env[63197]: INFO nova.compute.manager [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 919.254069] env[63197]: DEBUG oslo.service.loopingcall [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.254281] env[63197]: DEBUG nova.compute.manager [-] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.254370] env[63197]: DEBUG nova.network.neutron [-] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 919.363803] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.452113] env[63197]: DEBUG nova.network.neutron [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Added VIF to instance network info cache for port 9dfe2781-a394-4232-ac80-7e536c691f94. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3546}} [ 919.452113] env[63197]: DEBUG nova.network.neutron [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9dfe2781-a394-4232-ac80-7e536c691f94", "address": "fa:16:3e:b2:4f:a9", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dfe2781-a3", "ovs_interfaceid": "9dfe2781-a394-4232-ac80-7e536c691f94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.521075] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364282, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066548} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.521564] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.522847] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47dab55d-ff15-47b5-890f-dfa8f9fd509b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.530307] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1eb964eb-074b-42eb-bec9-c08d12248fc6 tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "232ea430-6a11-4dbc-9da9-0f734471d03b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.317s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.554819] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/52549788-b58f-4f5d-8443-a0c771099884.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.557472] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-819944f4-459c-4311-a571-c8be5518f668 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.582239] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364283, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.596036] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 919.596036] env[63197]: value = "task-1364284" [ 919.596036] env[63197]: _type = "Task" [ 919.596036] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.604696] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364284, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.632028] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.632028] env[63197]: DEBUG nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Instance network_info: |[{"id": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "address": "fa:16:3e:be:d6:04", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d53dd8b-1b", "ovs_interfaceid": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 919.632028] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] Acquired lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.632028] env[63197]: DEBUG nova.network.neutron [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Refreshing network info cache for port 2d53dd8b-1ba1-495d-99fc-265d17cb34a5 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 919.632433] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:d6:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d53dd8b-1ba1-495d-99fc-265d17cb34a5', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.640882] env[63197]: DEBUG oslo.service.loopingcall [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.645110] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 919.645387] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-210549b1-c379-4ba7-a710-a1a107d5f581 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.665935] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.665935] env[63197]: value = "task-1364285" [ 919.665935] env[63197]: _type = "Task" [ 919.665935] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.673691] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364285, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.954712] env[63197]: DEBUG oslo_concurrency.lockutils [req-4cd6de0d-e7c4-4ce9-b9cd-9b96356df053 req-4d2a48b2-aed1-4b14-a5ee-9cb28a68d107 service nova] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.955735] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.955735] env[63197]: DEBUG nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.024842] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 65df32fa-a6e7-4c3a-af8b-0422df8d9229 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.025131] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 9c1caa2b-c369-425a-8726-cddadf06f338 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.025367] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.025597] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.025812] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 638ef9c9-253b-4958-a660-6c1801408a51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.026221] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 180e0da7-f7ee-4fcd-be95-c2bf679278d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.026483] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 9d8c131d-1183-4508-ae2d-28e38a50e58d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.026735] env[63197]: WARNING nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a1f365de-b634-481a-b2e8-9bfc801a47f5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 920.026950] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance e87c0992-0c7c-4de8-ac68-800eb80cfbc4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.027185] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 52549788-b58f-4f5d-8443-a0c771099884 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.027414] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 537cc8cc-94e0-41b4-8204-6f995f2ea4d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.027699] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 920.027955] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 920.058743] env[63197]: DEBUG nova.network.neutron [-] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.083911] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364283, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549684} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.084741] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] e87c0992-0c7c-4de8-ac68-800eb80cfbc4/e87c0992-0c7c-4de8-ac68-800eb80cfbc4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.084955] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.085248] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f070761a-086b-4025-92fe-a7c8bff83311 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.092288] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 920.092288] env[63197]: value = "task-1364286" [ 920.092288] env[63197]: _type = "Task" [ 920.092288] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.099212] env[63197]: DEBUG nova.network.neutron [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updated VIF entry in instance network info cache for port 2d53dd8b-1ba1-495d-99fc-265d17cb34a5. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.099602] env[63197]: DEBUG nova.network.neutron [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance_info_cache with network_info: [{"id": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "address": "fa:16:3e:be:d6:04", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d53dd8b-1b", "ovs_interfaceid": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.115774] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.116219] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.178936] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364285, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.229754] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc1da06-aebf-45eb-b094-c7c1604aa166 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.237377] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027b3b03-27ae-4ea9-9319-05c9f0003674 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.267568] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2665d755-f862-4cba-87fd-6e292b64ab38 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.275631] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3df0e3-88ec-4f14-8ff8-e9b6f246c135 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.288822] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 920.493999] env[63197]: WARNING nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] 7ac08e03-1288-4a87-9a9c-257ba778c6d6 already exists in list: networks containing: ['7ac08e03-1288-4a87-9a9c-257ba778c6d6']. ignoring it [ 920.494239] env[63197]: WARNING nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] 7ac08e03-1288-4a87-9a9c-257ba778c6d6 already exists in list: networks containing: ['7ac08e03-1288-4a87-9a9c-257ba778c6d6']. ignoring it [ 920.494503] env[63197]: WARNING nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] 9dfe2781-a394-4232-ac80-7e536c691f94 already exists in list: port_ids containing: ['9dfe2781-a394-4232-ac80-7e536c691f94']. ignoring it [ 920.533905] env[63197]: DEBUG nova.compute.manager [req-80c56f27-441f-4351-8180-6385dda23596 req-e3e8f577-e280-4bff-b70d-56aa986a220f service nova] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Received event network-vif-deleted-1ad29d63-8ec8-498a-ad1d-7fd147ddc45a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.556013] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.556291] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.556477] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.556653] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.556813] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.558709] env[63197]: INFO nova.compute.manager [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Terminating instance [ 920.560384] env[63197]: DEBUG nova.compute.manager [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.560600] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.561422] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947d8d17-6ba0-4a0a-86e4-da1e09618d72 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.564472] env[63197]: INFO nova.compute.manager [-] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Took 1.31 seconds to deallocate network for instance. [ 920.570692] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.570922] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff7c2b1e-a57c-456e-94a5-ca3af2afb114 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.578762] env[63197]: DEBUG oslo_vmware.api [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 920.578762] env[63197]: value = "task-1364287" [ 920.578762] env[63197]: _type = "Task" [ 920.578762] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.588349] env[63197]: DEBUG oslo_vmware.api [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364287, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.602481] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113905} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.605797] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 920.606287] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d165c20-44d4-4443-b591-78ac7e589949 req-03b1a19d-b22a-4dd4-bf49-66080401c1a7 service nova] Releasing lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.607085] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5b6195-de97-4a0a-baa1-30afaed87375 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.637626] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] e87c0992-0c7c-4de8-ac68-800eb80cfbc4/e87c0992-0c7c-4de8-ac68-800eb80cfbc4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.637875] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364284, 'name': ReconfigVM_Task, 'duration_secs': 0.77323} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.638116] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f4b6683-763e-444d-aebf-686997849f34 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.652743] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/52549788-b58f-4f5d-8443-a0c771099884.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.653799] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26f73398-5506-433b-b8b0-6e952ea218ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.660328] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 920.660328] env[63197]: value = "task-1364289" [ 920.660328] env[63197]: _type = "Task" [ 920.660328] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.661737] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 920.661737] env[63197]: value = "task-1364288" [ 920.661737] env[63197]: _type = "Task" [ 920.661737] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.676313] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364289, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.679904] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364288, 'name': Rename_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.683944] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364285, 'name': CreateVM_Task, 'duration_secs': 0.576246} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.686073] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.686994] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.687171] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.687763] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.688031] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a77d7c5-1b73-42d8-aa30-28007bb357a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.692268] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 920.692268] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f541d1-71ba-6203-209d-b98ee3bdcf65" [ 920.692268] env[63197]: _type = "Task" [ 920.692268] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.699844] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f541d1-71ba-6203-209d-b98ee3bdcf65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.806793] env[63197]: DEBUG nova.network.neutron [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9dfe2781-a394-4232-ac80-7e536c691f94", "address": "fa:16:3e:b2:4f:a9", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dfe2781-a3", "ovs_interfaceid": "9dfe2781-a394-4232-ac80-7e536c691f94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.808775] env[63197]: ERROR nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [req-2218bb9a-c589-4921-beee-96964b2de48f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e9e01824-bb93-4102-9e5b-66d5dbc1990b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2218bb9a-c589-4921-beee-96964b2de48f"}]} [ 920.824317] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Refreshing inventories for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 920.839263] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Updating ProviderTree inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 920.839556] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 920.851354] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Refreshing aggregate associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, aggregates: None {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 920.872296] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Refreshing trait associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64 {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 921.011865] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e2b8b8-87e9-4093-9d0c-4b053d5a2cec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.019722] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e35e9ac-f377-4a8f-a8c8-952242fb3b2a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.049988] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f83fbd3-d65b-4caa-8b26-2d17ba075394 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.057366] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06f9d9a-213c-4a29-9152-419845de6948 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.070744] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 921.072664] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.087498] env[63197]: DEBUG oslo_vmware.api [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364287, 'name': PowerOffVM_Task, 'duration_secs': 0.198874} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.087743] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.087909] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.088198] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff246c9e-03f8-4e2b-a179-1b6c9cc33bc8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.146858] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.147107] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.147445] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Deleting the datastore file [datastore2] 65df32fa-a6e7-4c3a-af8b-0422df8d9229 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.147723] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1395c978-51e5-4406-bca1-f97d10733567 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.154490] env[63197]: DEBUG oslo_vmware.api [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for the task: (returnval){ [ 921.154490] env[63197]: value = "task-1364291" [ 921.154490] env[63197]: _type = "Task" [ 921.154490] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.162589] env[63197]: DEBUG oslo_vmware.api [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.172310] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364289, 'name': ReconfigVM_Task, 'duration_secs': 0.342073} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.172890] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Reconfigured VM instance instance-0000004c to attach disk [datastore1] e87c0992-0c7c-4de8-ac68-800eb80cfbc4/e87c0992-0c7c-4de8-ac68-800eb80cfbc4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.173526] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d87d9fa-cd01-43c5-9c19-420acff87411 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.177496] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364288, 'name': Rename_Task, 'duration_secs': 0.2018} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.178054] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.178277] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc8883a2-ce32-4b52-9ccb-d14c8924fe45 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.181927] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 921.181927] env[63197]: value = "task-1364292" [ 921.181927] env[63197]: _type = "Task" [ 921.181927] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.185625] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 921.185625] env[63197]: value = "task-1364293" [ 921.185625] env[63197]: _type = "Task" [ 921.185625] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.191318] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364292, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.198653] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364293, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.204163] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f541d1-71ba-6203-209d-b98ee3bdcf65, 'name': SearchDatastore_Task, 'duration_secs': 0.008712} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.204485] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.204715] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.204941] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.205099] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.205279] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.205517] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3bfd540a-9afc-4fd2-9f00-3b4b4bed4ee1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.212893] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.213103] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.213834] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7e32b3c-79e2-411a-9cad-a795decb69f9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.218888] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 921.218888] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5295cf44-787c-3897-3f3d-ac4922604eab" [ 921.218888] env[63197]: _type = "Task" [ 921.218888] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.226688] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5295cf44-787c-3897-3f3d-ac4922604eab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.313085] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.313440] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.313620] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.314679] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5b3273-7a70-4cf5-af65-2b1a6ecad574 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.332231] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 921.332520] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 921.332677] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 921.332855] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 921.333013] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 921.333238] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 921.333460] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 921.333636] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 921.333804] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 921.333979] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 921.334179] env[63197]: DEBUG nova.virt.hardware [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 921.340529] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfiguring VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 921.340867] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75240574-ec33-4e20-871a-22013ec9a1eb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.357411] env[63197]: DEBUG oslo_vmware.api [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 921.357411] env[63197]: value = "task-1364294" [ 921.357411] env[63197]: _type = "Task" [ 921.357411] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.367364] env[63197]: DEBUG oslo_vmware.api [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364294, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.612766] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Updated inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with generation 86 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 921.613170] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Updating resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b generation from 86 to 87 during operation: update_inventory {{(pid=63197) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 921.613408] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 921.664529] env[63197]: DEBUG oslo_vmware.api [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Task: {'id': task-1364291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154041} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.664827] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.664962] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.665155] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.665370] env[63197]: INFO nova.compute.manager [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Took 1.10 seconds to destroy the instance on the hypervisor. [ 921.665610] env[63197]: DEBUG oslo.service.loopingcall [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.666120] env[63197]: DEBUG nova.compute.manager [-] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 921.666219] env[63197]: DEBUG nova.network.neutron [-] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 921.694431] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364292, 'name': Rename_Task, 'duration_secs': 0.139552} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.697377] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.697649] env[63197]: DEBUG oslo_vmware.api [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364293, 'name': PowerOnVM_Task, 'duration_secs': 0.439557} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.697848] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7fce61f-a86f-4354-95c7-87aa047f45fd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.699432] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.699744] env[63197]: INFO nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Took 7.58 seconds to spawn the instance on the hypervisor. [ 921.699807] env[63197]: DEBUG nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.700846] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0402ca-972c-4f35-a170-d175e28cbe3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.710552] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 921.710552] env[63197]: value = "task-1364295" [ 921.710552] env[63197]: _type = "Task" [ 921.710552] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.718176] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364295, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.728678] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5295cf44-787c-3897-3f3d-ac4922604eab, 'name': SearchDatastore_Task, 'duration_secs': 0.008666} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.731397] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01b36976-3a3b-4f6d-a0b7-c5b205b72171 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.736741] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 921.736741] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526bd5a2-c0e3-c090-81f7-2e0c5e7e6b10" [ 921.736741] env[63197]: _type = "Task" [ 921.736741] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.744994] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526bd5a2-c0e3-c090-81f7-2e0c5e7e6b10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.868263] env[63197]: DEBUG oslo_vmware.api [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364294, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.120542] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 922.120772] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.141s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.121093] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.758s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.121321] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.124119] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.051s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.124378] env[63197]: DEBUG nova.objects.instance [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lazy-loading 'resources' on Instance uuid 9d8c131d-1183-4508-ae2d-28e38a50e58d {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.148334] env[63197]: INFO nova.scheduler.client.report [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted allocations for instance a1f365de-b634-481a-b2e8-9bfc801a47f5 [ 922.221905] env[63197]: INFO nova.compute.manager [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Took 19.65 seconds to build instance. [ 922.227412] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364295, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.247172] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526bd5a2-c0e3-c090-81f7-2e0c5e7e6b10, 'name': SearchDatastore_Task, 'duration_secs': 0.00968} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.247455] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.247788] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 537cc8cc-94e0-41b4-8204-6f995f2ea4d3/537cc8cc-94e0-41b4-8204-6f995f2ea4d3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.248065] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0405c830-935f-4868-8c02-a8e6787498d9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.255104] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 922.255104] env[63197]: value = "task-1364296" [ 922.255104] env[63197]: _type = "Task" [ 922.255104] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.263770] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364296, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.369671] env[63197]: DEBUG oslo_vmware.api [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364294, 'name': ReconfigVM_Task, 'duration_secs': 0.573984} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.370231] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.370460] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfigured VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 922.415424] env[63197]: DEBUG nova.network.neutron [-] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.558455] env[63197]: DEBUG nova.compute.manager [req-e99a92ce-59e9-48c3-8974-5e95c2e9c4e2 req-cca3ad50-b2ff-4343-a8f6-04f252dd3e42 service nova] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Received event network-vif-deleted-5f405bfb-7d3e-4876-8fe0-2dee955c3325 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.657769] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d8227ff2-e70a-4f40-9d31-6fc4674dda62 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "a1f365de-b634-481a-b2e8-9bfc801a47f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.322s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.726280] env[63197]: DEBUG oslo_vmware.api [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364295, 'name': PowerOnVM_Task, 'duration_secs': 0.979821} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.726615] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 922.726773] env[63197]: INFO nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Took 14.71 seconds to spawn the instance on the hypervisor. [ 922.726947] env[63197]: DEBUG nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.727394] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6317d2f7-5c20-4726-8abb-2520c890049d tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "52549788-b58f-4f5d-8443-a0c771099884" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.162s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.728145] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e8087c7-5c9e-4df7-8c08-a15d97bd25c8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.766723] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364296, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481093} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.767018] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 537cc8cc-94e0-41b4-8204-6f995f2ea4d3/537cc8cc-94e0-41b4-8204-6f995f2ea4d3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.767312] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.767793] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f06fb43d-b57d-43a6-840a-017e84fd28fb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.776654] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 922.776654] env[63197]: value = "task-1364297" [ 922.776654] env[63197]: _type = "Task" [ 922.776654] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.788235] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364297, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.788535] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44617883-5d9a-4501-800b-70962b0e2224 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.796953] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cf96c2-6e25-4498-931e-0db32fffcfe3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.833517] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233505a4-42d7-4343-8971-660a7e34013d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.845762] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b2099e-ec7e-45cf-8aa8-ffd5e2799b3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.860828] env[63197]: DEBUG nova.compute.provider_tree [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.876582] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7bde630a-f6f2-4a13-8690-7e38b91d25df tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.448s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.918097] env[63197]: INFO nova.compute.manager [-] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Took 1.25 seconds to deallocate network for instance. [ 923.249588] env[63197]: INFO nova.compute.manager [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Took 35.77 seconds to build instance. [ 923.290251] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364297, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062765} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.290717] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.291983] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be69141d-c95f-447d-951c-6503950206d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.316032] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 537cc8cc-94e0-41b4-8204-6f995f2ea4d3/537cc8cc-94e0-41b4-8204-6f995f2ea4d3.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.316431] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-224e289c-3785-48c9-a854-1d8b085ae9a2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.337567] env[63197]: INFO nova.compute.manager [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Rescuing [ 923.337827] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.338045] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.338152] env[63197]: DEBUG nova.network.neutron [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 923.339460] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 923.339460] env[63197]: value = "task-1364298" [ 923.339460] env[63197]: _type = "Task" [ 923.339460] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.349376] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364298, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.381873] env[63197]: ERROR nova.scheduler.client.report [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] [req-c22df570-3102-48b7-9067-41ed50c1dbbc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e9e01824-bb93-4102-9e5b-66d5dbc1990b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c22df570-3102-48b7-9067-41ed50c1dbbc"}]} [ 923.397218] env[63197]: DEBUG nova.scheduler.client.report [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Refreshing inventories for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 923.409896] env[63197]: DEBUG nova.scheduler.client.report [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Updating ProviderTree inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 923.410175] env[63197]: DEBUG nova.compute.provider_tree [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.421204] env[63197]: DEBUG nova.scheduler.client.report [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Refreshing aggregate associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, aggregates: None {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 923.426279] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.437859] env[63197]: DEBUG nova.scheduler.client.report [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Refreshing trait associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64 {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 923.585034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0e09c0-fe5f-4898-8268-857a570471aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.592097] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b34c1d5-c386-4826-95ce-b5fd86f995f5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.625309] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2f53b6-d5e4-44e4-a32b-dbd5532861f8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.635975] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6e4e5e-ed1c-498a-a7db-99aa9734f89c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.650692] env[63197]: DEBUG nova.compute.provider_tree [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 923.752146] env[63197]: DEBUG oslo_concurrency.lockutils [None req-168c98b5-f5ee-44dc-a46c-a82f44913d72 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.516s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.814338] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "4757f91b-21f1-4a16-bda3-729b12ddf86c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.815210] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.849445] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364298, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.131355] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.131606] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.131801] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.131979] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.132238] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.134264] env[63197]: INFO nova.compute.manager [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Terminating instance [ 924.136231] env[63197]: DEBUG nova.compute.manager [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 924.136417] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.137246] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6216acc0-6f0a-4d41-86d9-b959326f4e66 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.146129] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.146389] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03a17eb2-b86e-4e32-9b99-28dd6dd75f18 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.149025] env[63197]: DEBUG nova.network.neutron [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Updating instance_info_cache with network_info: [{"id": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "address": "fa:16:3e:b8:5c:89", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd634e1d-f1", "ovs_interfaceid": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.156386] env[63197]: DEBUG oslo_vmware.api [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 924.156386] env[63197]: value = "task-1364299" [ 924.156386] env[63197]: _type = "Task" [ 924.156386] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.165770] env[63197]: DEBUG oslo_vmware.api [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.186917] env[63197]: DEBUG nova.scheduler.client.report [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Updated inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 924.187289] env[63197]: DEBUG nova.compute.provider_tree [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Updating resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b generation from 88 to 89 during operation: update_inventory {{(pid=63197) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 924.187545] env[63197]: DEBUG nova.compute.provider_tree [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 924.317411] env[63197]: DEBUG nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.351489] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364298, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.625670] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-33a215bc-a3d4-48cf-8856-1da55204dbdf" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.625915] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-33a215bc-a3d4-48cf-8856-1da55204dbdf" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.626377] env[63197]: DEBUG nova.objects.instance [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'flavor' on Instance uuid 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.651117] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.667122] env[63197]: DEBUG oslo_vmware.api [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364299, 'name': PowerOffVM_Task, 'duration_secs': 0.265689} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.667391] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.667553] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.669417] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d50e66b-b673-4e3d-81eb-d6f551a34f36 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.692086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.568s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.694667] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.269s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.694902] env[63197]: DEBUG nova.objects.instance [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lazy-loading 'resources' on Instance uuid 65df32fa-a6e7-4c3a-af8b-0422df8d9229 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.717133] env[63197]: INFO nova.scheduler.client.report [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Deleted allocations for instance 9d8c131d-1183-4508-ae2d-28e38a50e58d [ 924.836631] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.840260] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.840530] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.840748] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Deleting the datastore file [datastore1] e87c0992-0c7c-4de8-ac68-800eb80cfbc4 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.841049] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4235e4f4-c68c-481f-918c-e8753d778881 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.851822] env[63197]: DEBUG oslo_vmware.api [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 924.851822] env[63197]: value = "task-1364301" [ 924.851822] env[63197]: _type = "Task" [ 924.851822] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.855026] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364298, 'name': ReconfigVM_Task, 'duration_secs': 1.330781} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.857949] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 537cc8cc-94e0-41b4-8204-6f995f2ea4d3/537cc8cc-94e0-41b4-8204-6f995f2ea4d3.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.858582] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce001f69-7d10-4bfe-adf5-a0910bcb9ca4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.864353] env[63197]: DEBUG oslo_vmware.api [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364301, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.865511] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 924.865511] env[63197]: value = "task-1364302" [ 924.865511] env[63197]: _type = "Task" [ 924.865511] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.872888] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364302, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.181943] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.182254] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1092c94b-5ea4-45a4-9feb-0865a24c713f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.191595] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 925.191595] env[63197]: value = "task-1364303" [ 925.191595] env[63197]: _type = "Task" [ 925.191595] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.202287] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.224593] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a704a296-6156-440f-bd8c-151848f811c1 tempest-ImagesTestJSON-1924325623 tempest-ImagesTestJSON-1924325623-project-member] Lock "9d8c131d-1183-4508-ae2d-28e38a50e58d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.121s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.265150] env[63197]: DEBUG nova.objects.instance [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'pci_requests' on Instance uuid 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.335204] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5625e21-420f-4a1f-9233-8e6959bf6236 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.342813] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0736116-7633-4c7d-9ddd-c591c868da73 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.379839] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d633706b-1575-4cd2-80a2-f4c84ed23753 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.386976] env[63197]: DEBUG oslo_vmware.api [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364301, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171725} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.391782] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.392185] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.392222] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.392398] env[63197]: INFO nova.compute.manager [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Took 1.26 seconds to destroy the instance on the hypervisor. [ 925.392650] env[63197]: DEBUG oslo.service.loopingcall [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.392872] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364302, 'name': Rename_Task, 'duration_secs': 0.158449} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.393174] env[63197]: DEBUG nova.compute.manager [-] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.393296] env[63197]: DEBUG nova.network.neutron [-] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.394948] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.396151] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4193988b-0f6d-4925-a0e3-b38f4535d1ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.400274] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75c1ec41-45b5-4030-9dd0-18c208d55cec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.406505] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 925.406505] env[63197]: value = "task-1364304" [ 925.406505] env[63197]: _type = "Task" [ 925.406505] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.414109] env[63197]: DEBUG nova.compute.provider_tree [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.423526] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364304, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.702980] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364303, 'name': PowerOffVM_Task, 'duration_secs': 0.203616} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.703351] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.704212] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0378817-dac2-4cb4-8719-cef17bbd0f19 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.723214] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa93360c-dfff-47d2-bb5c-c8f2899cc60d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.755948] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.756260] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a4958d6-c860-470e-aad6-8f50bf8e8bee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.763504] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 925.763504] env[63197]: value = "task-1364305" [ 925.763504] env[63197]: _type = "Task" [ 925.763504] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.774024] env[63197]: DEBUG nova.objects.base [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Object Instance<9c1caa2b-c369-425a-8726-cddadf06f338> lazy-loaded attributes: flavor,pci_requests {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 925.775015] env[63197]: DEBUG nova.network.neutron [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 925.777380] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.789759] env[63197]: DEBUG nova.compute.manager [req-ad263fcd-9e24-46d8-8685-674bc89e7c99 req-3187dfe2-9216-45aa-b7f6-8295fc5c20e9 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-vif-deleted-4403e435-3845-48ec-a898-a1a932fea8d3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.790073] env[63197]: INFO nova.compute.manager [req-ad263fcd-9e24-46d8-8685-674bc89e7c99 req-3187dfe2-9216-45aa-b7f6-8295fc5c20e9 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Neutron deleted interface 4403e435-3845-48ec-a898-a1a932fea8d3; detaching it from the instance and deleting it from the info cache [ 925.790427] env[63197]: DEBUG nova.network.neutron [req-ad263fcd-9e24-46d8-8685-674bc89e7c99 req-3187dfe2-9216-45aa-b7f6-8295fc5c20e9 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Updating instance_info_cache with network_info: [{"id": "cf42dede-8506-413b-a119-cbe31f3be836", "address": "fa:16:3e:2e:81:9f", "network": {"id": "1fe2f14b-d6c0-42f3-a4e1-f70a3cf56e6d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1080454701", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.102", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c1520c99-af74-4d61-a8ae-56aef56ef4f0", "external-id": "nsx-vlan-transportzone-891", "segmentation_id": 891, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42dede-85", "ovs_interfaceid": "cf42dede-8506-413b-a119-cbe31f3be836", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "address": "fa:16:3e:cb:2b:5e", "network": {"id": "f63fd958-33f7-4275-ad22-bb6545390ccc", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1942842288", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc856ee9c-0a", "ovs_interfaceid": "c856ee9c-0a79-46b8-a8ec-f6283349a05e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.854453] env[63197]: DEBUG nova.policy [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 925.923023] env[63197]: DEBUG nova.scheduler.client.report [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.930726] env[63197]: DEBUG oslo_vmware.api [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364304, 'name': PowerOnVM_Task, 'duration_secs': 0.467283} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.930726] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.930726] env[63197]: INFO nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Took 8.53 seconds to spawn the instance on the hypervisor. [ 925.930726] env[63197]: DEBUG nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.930726] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323b75e6-d682-4b31-be1e-bff8743ddd49 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.273642] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] VM already powered off {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 926.273934] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 926.274113] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.274268] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.274452] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.274687] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8daf642b-e24c-4d9c-b3aa-0a2f916cf680 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.283222] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.283411] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 926.284130] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f607b2b-328e-425c-86f6-e0f0ee503b48 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.289404] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 926.289404] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526b80f2-2945-21a4-fb52-4c626e9aa45a" [ 926.289404] env[63197]: _type = "Task" [ 926.289404] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.293065] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24051f89-6c74-4d6d-b7f3-09bc25d160c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.302762] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526b80f2-2945-21a4-fb52-4c626e9aa45a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.305558] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2550407-af2d-4ec3-a8b0-27639f692bff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.335154] env[63197]: DEBUG nova.compute.manager [req-ad263fcd-9e24-46d8-8685-674bc89e7c99 req-3187dfe2-9216-45aa-b7f6-8295fc5c20e9 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Detach interface failed, port_id=4403e435-3845-48ec-a898-a1a932fea8d3, reason: Instance e87c0992-0c7c-4de8-ac68-800eb80cfbc4 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 926.425088] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.427594] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.591s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.429104] env[63197]: INFO nova.compute.claims [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.448387] env[63197]: INFO nova.compute.manager [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Took 18.57 seconds to build instance. [ 926.461285] env[63197]: INFO nova.scheduler.client.report [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Deleted allocations for instance 65df32fa-a6e7-4c3a-af8b-0422df8d9229 [ 926.556820] env[63197]: DEBUG nova.network.neutron [-] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.801680] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526b80f2-2945-21a4-fb52-4c626e9aa45a, 'name': SearchDatastore_Task, 'duration_secs': 0.009006} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.802640] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c800ee6-5f48-4142-a0af-36f4af8749ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.808322] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 926.808322] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524fd18c-4db0-e5b8-6432-31ec82bf7148" [ 926.808322] env[63197]: _type = "Task" [ 926.808322] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.817679] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524fd18c-4db0-e5b8-6432-31ec82bf7148, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.950372] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3b2b1a2-09f2-41ad-9179-03044b7298e7 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.088s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.976026] env[63197]: DEBUG oslo_concurrency.lockutils [None req-72b2d98a-8d4b-41b1-9b81-c2944edae01c tempest-FloatingIPsAssociationTestJSON-689624108 tempest-FloatingIPsAssociationTestJSON-689624108-project-member] Lock "65df32fa-a6e7-4c3a-af8b-0422df8d9229" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.417s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.059667] env[63197]: INFO nova.compute.manager [-] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Took 1.67 seconds to deallocate network for instance. [ 927.318744] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524fd18c-4db0-e5b8-6432-31ec82bf7148, 'name': SearchDatastore_Task, 'duration_secs': 0.009693} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.319527] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.319801] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk. {{(pid=63197) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 927.320075] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f160d52e-1d77-49c7-8a00-443119cace46 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.327244] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 927.327244] env[63197]: value = "task-1364306" [ 927.327244] env[63197]: _type = "Task" [ 927.327244] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.335144] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364306, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.568164] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.642412] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86601d1-01f3-4cef-9837-b73f7c91ab9a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.654942] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed3d8b4-d741-428c-a897-470f489a3b5a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.688290] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18652c4a-904a-4351-bb57-353a3c235fc6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.698109] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7bcdf3-15dc-43e5-8524-d8929fce8fb1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.717094] env[63197]: DEBUG nova.compute.provider_tree [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.798472] env[63197]: DEBUG nova.network.neutron [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Successfully updated port: 33a215bc-a3d4-48cf-8856-1da55204dbdf {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.844790] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364306, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485063} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.844790] env[63197]: INFO nova.virt.vmwareapi.ds_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk. [ 927.846654] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c473ff5-5be5-45f4-a78c-e9c4883cb626 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.879363] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.880778] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84be70f7-5428-494e-b5c0-d8ff6c92d8c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.895718] env[63197]: DEBUG nova.compute.manager [req-a82c6cdf-c898-4f2c-bdf1-e7aa73eebf5f req-d6cbd783-9747-494a-acd0-66c961d6a44a service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-vif-plugged-33a215bc-a3d4-48cf-8856-1da55204dbdf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.895718] env[63197]: DEBUG oslo_concurrency.lockutils [req-a82c6cdf-c898-4f2c-bdf1-e7aa73eebf5f req-d6cbd783-9747-494a-acd0-66c961d6a44a service nova] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.895718] env[63197]: DEBUG oslo_concurrency.lockutils [req-a82c6cdf-c898-4f2c-bdf1-e7aa73eebf5f req-d6cbd783-9747-494a-acd0-66c961d6a44a service nova] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.895718] env[63197]: DEBUG oslo_concurrency.lockutils [req-a82c6cdf-c898-4f2c-bdf1-e7aa73eebf5f req-d6cbd783-9747-494a-acd0-66c961d6a44a service nova] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.895718] env[63197]: DEBUG nova.compute.manager [req-a82c6cdf-c898-4f2c-bdf1-e7aa73eebf5f req-d6cbd783-9747-494a-acd0-66c961d6a44a service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] No waiting events found dispatching network-vif-plugged-33a215bc-a3d4-48cf-8856-1da55204dbdf {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 927.895957] env[63197]: WARNING nova.compute.manager [req-a82c6cdf-c898-4f2c-bdf1-e7aa73eebf5f req-d6cbd783-9747-494a-acd0-66c961d6a44a service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received unexpected event network-vif-plugged-33a215bc-a3d4-48cf-8856-1da55204dbdf for instance with vm_state active and task_state None. [ 927.902315] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 927.902315] env[63197]: value = "task-1364307" [ 927.902315] env[63197]: _type = "Task" [ 927.902315] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.910732] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364307, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.028192] env[63197]: DEBUG nova.compute.manager [req-f25ffca7-230f-48e0-976a-5af5d6446c05 req-e5b31896-8c57-4ae5-8464-e67b21404701 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-vif-deleted-c856ee9c-0a79-46b8-a8ec-f6283349a05e {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.028473] env[63197]: DEBUG nova.compute.manager [req-f25ffca7-230f-48e0-976a-5af5d6446c05 req-e5b31896-8c57-4ae5-8464-e67b21404701 service nova] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Received event network-vif-deleted-cf42dede-8506-413b-a119-cbe31f3be836 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.223493] env[63197]: DEBUG nova.scheduler.client.report [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.304147] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.304321] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.307047] env[63197]: DEBUG nova.network.neutron [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 928.413714] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364307, 'name': ReconfigVM_Task, 'duration_secs': 0.278949} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.414108] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 52549788-b58f-4f5d-8443-a0c771099884/f760fedc-0b5b-4c56-acbe-239b47c945e8-rescue.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.414838] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5128aad2-b921-4424-9671-763da75df983 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.441994] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05283119-d20d-4e59-83ed-5b8158361947 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.459558] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 928.459558] env[63197]: value = "task-1364308" [ 928.459558] env[63197]: _type = "Task" [ 928.459558] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.467960] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364308, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.729356] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.729869] env[63197]: DEBUG nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.734470] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.169s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.734797] env[63197]: DEBUG nova.objects.instance [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lazy-loading 'resources' on Instance uuid e87c0992-0c7c-4de8-ac68-800eb80cfbc4 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.855128] env[63197]: WARNING nova.network.neutron [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] 7ac08e03-1288-4a87-9a9c-257ba778c6d6 already exists in list: networks containing: ['7ac08e03-1288-4a87-9a9c-257ba778c6d6']. ignoring it [ 928.855128] env[63197]: WARNING nova.network.neutron [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] 7ac08e03-1288-4a87-9a9c-257ba778c6d6 already exists in list: networks containing: ['7ac08e03-1288-4a87-9a9c-257ba778c6d6']. ignoring it [ 928.970937] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364308, 'name': ReconfigVM_Task, 'duration_secs': 0.143878} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.971306] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.971551] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6c573d2-792b-4325-b6f2-e610ea257fef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.977932] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 928.977932] env[63197]: value = "task-1364309" [ 928.977932] env[63197]: _type = "Task" [ 928.977932] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.985581] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.238902] env[63197]: DEBUG nova.compute.utils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.247323] env[63197]: DEBUG nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 929.247504] env[63197]: DEBUG nova.network.neutron [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 929.268662] env[63197]: DEBUG nova.compute.manager [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Stashing vm_state: active {{(pid=63197) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 929.294256] env[63197]: DEBUG nova.policy [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12d00044b65c48f484ac9fea595d6179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80138785cfdb4a2188ffb972da437646', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 929.445017] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f8bea7-799e-4f1a-96f5-79d5d0d0bf23 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.455951] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e206896-6381-49d8-afe0-e4b124d5a27a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.501643] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee2d58d-dde1-45fc-80fe-4adfb026a171 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.511731] env[63197]: DEBUG oslo_vmware.api [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364309, 'name': PowerOnVM_Task, 'duration_secs': 0.428772} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.514480] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.518278] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dd27cb-d60f-4d10-927f-f99b7efc8d97 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.526630] env[63197]: DEBUG nova.compute.manager [None req-0c18d37d-f40b-4578-9eae-4cc1bacab5d5 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.526630] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f79125-eaee-428c-a441-29daad381038 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.542562] env[63197]: DEBUG nova.compute.provider_tree [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.559395] env[63197]: DEBUG nova.network.neutron [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Successfully created port: 4bebe341-10f9-4ff8-a0b7-3312eb20074f {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.747887] env[63197]: DEBUG nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.766286] env[63197]: DEBUG nova.network.neutron [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9dfe2781-a394-4232-ac80-7e536c691f94", "address": "fa:16:3e:b2:4f:a9", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dfe2781-a3", "ovs_interfaceid": "9dfe2781-a394-4232-ac80-7e536c691f94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "33a215bc-a3d4-48cf-8856-1da55204dbdf", "address": "fa:16:3e:ba:ff:04", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a215bc-a3", "ovs_interfaceid": "33a215bc-a3d4-48cf-8856-1da55204dbdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.795334] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.048624] env[63197]: DEBUG nova.scheduler.client.report [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.183080] env[63197]: DEBUG nova.compute.manager [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-changed-33a215bc-a3d4-48cf-8856-1da55204dbdf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.183186] env[63197]: DEBUG nova.compute.manager [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing instance network info cache due to event network-changed-33a215bc-a3d4-48cf-8856-1da55204dbdf. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 930.183400] env[63197]: DEBUG oslo_concurrency.lockutils [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.272799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.272799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.272799] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.272799] env[63197]: DEBUG oslo_concurrency.lockutils [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.273442] env[63197]: DEBUG nova.network.neutron [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Refreshing network info cache for port 33a215bc-a3d4-48cf-8856-1da55204dbdf {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 930.278847] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5d9e1b-e820-4f5d-9904-c91afcf35a7d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.302683] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.302947] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.303101] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.303307] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.303453] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.303594] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.303792] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.303978] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.304113] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.304274] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.304439] env[63197]: DEBUG nova.virt.hardware [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.311512] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfiguring VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 930.315019] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0ba7490-b2e3-412d-852c-9009990e22cd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.331670] env[63197]: DEBUG oslo_vmware.api [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 930.331670] env[63197]: value = "task-1364310" [ 930.331670] env[63197]: _type = "Task" [ 930.331670] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.343750] env[63197]: DEBUG oslo_vmware.api [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364310, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.454265] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.454265] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.556592] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.819s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.557354] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.761s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.587252] env[63197]: INFO nova.scheduler.client.report [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Deleted allocations for instance e87c0992-0c7c-4de8-ac68-800eb80cfbc4 [ 930.667455] env[63197]: INFO nova.compute.manager [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Unrescuing [ 930.667727] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.667879] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquired lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.668052] env[63197]: DEBUG nova.network.neutron [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.756887] env[63197]: DEBUG nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.792879] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.793388] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.793630] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.794843] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.795294] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.795477] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.795995] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.796252] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.796597] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.796924] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.797407] env[63197]: DEBUG nova.virt.hardware [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.799317] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c3dbdf-a662-4a76-863b-11de612021a7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.814244] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352f807d-9395-4069-822b-e1bb86e16e8e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.844557] env[63197]: DEBUG oslo_vmware.api [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364310, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.958565] env[63197]: DEBUG nova.compute.utils [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.069132] env[63197]: INFO nova.compute.claims [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.098536] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b266068c-9e78-44a7-a469-6930b40b995d tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "e87c0992-0c7c-4de8-ac68-800eb80cfbc4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.967s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.332076] env[63197]: DEBUG nova.network.neutron [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Successfully updated port: 4bebe341-10f9-4ff8-a0b7-3312eb20074f {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 931.350604] env[63197]: DEBUG oslo_vmware.api [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364310, 'name': ReconfigVM_Task, 'duration_secs': 0.918497} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.351777] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.352127] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfigured VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 931.391485] env[63197]: DEBUG nova.network.neutron [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updated VIF entry in instance network info cache for port 33a215bc-a3d4-48cf-8856-1da55204dbdf. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 931.391938] env[63197]: DEBUG nova.network.neutron [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9dfe2781-a394-4232-ac80-7e536c691f94", "address": "fa:16:3e:b2:4f:a9", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dfe2781-a3", "ovs_interfaceid": "9dfe2781-a394-4232-ac80-7e536c691f94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "33a215bc-a3d4-48cf-8856-1da55204dbdf", "address": "fa:16:3e:ba:ff:04", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a215bc-a3", "ovs_interfaceid": "33a215bc-a3d4-48cf-8856-1da55204dbdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.440396] env[63197]: DEBUG nova.network.neutron [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Updating instance_info_cache with network_info: [{"id": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "address": "fa:16:3e:b8:5c:89", "network": {"id": "ede3f7fd-abf8-4bcf-bddf-c40bee22b62e", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-50283780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0d992f8d3d334b8783f3dc5eff1cf423", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089ef678-58b4-4bf0-a39d-b94b2d364291", "external-id": "nsx-vlan-transportzone-675", "segmentation_id": 675, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd634e1d-f1", "ovs_interfaceid": "dd634e1d-f13f-4709-a49b-4a669d1286f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.462155] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.580293] env[63197]: INFO nova.compute.resource_tracker [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating resource usage from migration 90c18999-4dd5-4787-8836-75a6a0374213 [ 931.750019] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317a27db-d13b-4283-b516-a5062a42b403 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.756442] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c44988f-dd78-4a09-8ea8-6faff3dba3de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.788767] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d284e3-0805-4a36-96e8-faa0791f9f96 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.797569] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d7fca3-35b3-483a-a72b-6683c9e0ae9b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.812318] env[63197]: DEBUG nova.compute.provider_tree [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.840934] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-4757f91b-21f1-4a16-bda3-729b12ddf86c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.840934] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-4757f91b-21f1-4a16-bda3-729b12ddf86c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.840934] env[63197]: DEBUG nova.network.neutron [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 931.857670] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f0da079-2b95-442a-81cd-b27000ce4bc6 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-33a215bc-a3d4-48cf-8856-1da55204dbdf" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.231s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.895515] env[63197]: DEBUG oslo_concurrency.lockutils [req-c21941e7-ab6c-4f77-82cd-11deaec26826 req-232154a5-df49-446c-b66c-e6a3e105a2a0 service nova] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.945309] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Releasing lock "refresh_cache-52549788-b58f-4f5d-8443-a0c771099884" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.945309] env[63197]: DEBUG nova.objects.instance [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lazy-loading 'flavor' on Instance uuid 52549788-b58f-4f5d-8443-a0c771099884 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.316522] env[63197]: DEBUG nova.scheduler.client.report [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.368775] env[63197]: DEBUG nova.compute.manager [req-c260e6f7-d6cb-4895-bf78-7bbc25d05efa req-c7adc8ae-1393-40e5-9b80-0807e5059ab3 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Received event network-vif-plugged-4bebe341-10f9-4ff8-a0b7-3312eb20074f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.369082] env[63197]: DEBUG oslo_concurrency.lockutils [req-c260e6f7-d6cb-4895-bf78-7bbc25d05efa req-c7adc8ae-1393-40e5-9b80-0807e5059ab3 service nova] Acquiring lock "4757f91b-21f1-4a16-bda3-729b12ddf86c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.369208] env[63197]: DEBUG oslo_concurrency.lockutils [req-c260e6f7-d6cb-4895-bf78-7bbc25d05efa req-c7adc8ae-1393-40e5-9b80-0807e5059ab3 service nova] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.369402] env[63197]: DEBUG oslo_concurrency.lockutils [req-c260e6f7-d6cb-4895-bf78-7bbc25d05efa req-c7adc8ae-1393-40e5-9b80-0807e5059ab3 service nova] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.369569] env[63197]: DEBUG nova.compute.manager [req-c260e6f7-d6cb-4895-bf78-7bbc25d05efa req-c7adc8ae-1393-40e5-9b80-0807e5059ab3 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] No waiting events found dispatching network-vif-plugged-4bebe341-10f9-4ff8-a0b7-3312eb20074f {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.369756] env[63197]: WARNING nova.compute.manager [req-c260e6f7-d6cb-4895-bf78-7bbc25d05efa req-c7adc8ae-1393-40e5-9b80-0807e5059ab3 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Received unexpected event network-vif-plugged-4bebe341-10f9-4ff8-a0b7-3312eb20074f for instance with vm_state building and task_state spawning. [ 932.394071] env[63197]: DEBUG nova.network.neutron [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.454021] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9433791f-3567-422a-9644-24790e2448f5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.479707] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.481451] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-116c6d70-d169-40b8-916c-5ffcdf0f2ba2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.488345] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 932.488345] env[63197]: value = "task-1364311" [ 932.488345] env[63197]: _type = "Task" [ 932.488345] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.499481] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364311, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.551367] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.551367] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.551367] env[63197]: INFO nova.compute.manager [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Attaching volume c29df051-5e11-4357-9d18-1c7dc22c901b to /dev/sdb [ 932.593013] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9ccf3c-3bd1-4fc1-bb95-7a03ddbede59 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.602568] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9ac766-e742-450b-91cf-677335e05749 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.617125] env[63197]: DEBUG nova.virt.block_device [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updating existing volume attachment record: 36911a23-1fec-4829-9c29-07fd6f709662 {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 932.661540] env[63197]: DEBUG nova.network.neutron [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Updating instance_info_cache with network_info: [{"id": "4bebe341-10f9-4ff8-a0b7-3312eb20074f", "address": "fa:16:3e:70:fb:da", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bebe341-10", "ovs_interfaceid": "4bebe341-10f9-4ff8-a0b7-3312eb20074f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.822815] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.266s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.822815] env[63197]: INFO nova.compute.manager [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Migrating [ 932.822815] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.822815] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.999188] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364311, 'name': PowerOffVM_Task, 'duration_secs': 0.338358} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.999506] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.005915] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 933.006270] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34f03123-3faf-40fc-8fb4-9b574da7d236 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.033120] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 933.033120] env[63197]: value = "task-1364315" [ 933.033120] env[63197]: _type = "Task" [ 933.033120] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.042202] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364315, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.167033] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-4757f91b-21f1-4a16-bda3-729b12ddf86c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.167033] env[63197]: DEBUG nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Instance network_info: |[{"id": "4bebe341-10f9-4ff8-a0b7-3312eb20074f", "address": "fa:16:3e:70:fb:da", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bebe341-10", "ovs_interfaceid": "4bebe341-10f9-4ff8-a0b7-3312eb20074f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 933.167033] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:fb:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4bebe341-10f9-4ff8-a0b7-3312eb20074f', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.178642] env[63197]: DEBUG oslo.service.loopingcall [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.179204] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.179741] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b73341aa-5c03-48ab-b9b7-064c21c4c0d8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.207384] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.207384] env[63197]: value = "task-1364316" [ 933.207384] env[63197]: _type = "Task" [ 933.207384] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.215861] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364316, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.329952] env[63197]: INFO nova.compute.rpcapi [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 933.329952] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.547031] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364315, 'name': ReconfigVM_Task, 'duration_secs': 0.214} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.547764] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.548242] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.548708] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84937547-463a-40d6-a817-7cc91790dd2d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.557064] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 933.557064] env[63197]: value = "task-1364317" [ 933.557064] env[63197]: _type = "Task" [ 933.557064] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.565853] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364317, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.605301] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "4d12bac4-da42-42bd-9361-7015f3be9693" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.605626] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "4d12bac4-da42-42bd-9361-7015f3be9693" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.723051] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364316, 'name': CreateVM_Task, 'duration_secs': 0.43051} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.723051] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 933.723051] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.723051] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.723051] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.723051] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55ef18ad-bdbf-4c16-84df-8d72e475c527 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.728858] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 933.728858] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52783286-fabf-4fea-de21-8ce8a90f569e" [ 933.728858] env[63197]: _type = "Task" [ 933.728858] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.740663] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52783286-fabf-4fea-de21-8ce8a90f569e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.850908] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.850908] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.850908] env[63197]: DEBUG nova.network.neutron [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 933.876664] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-9dfe2781-a394-4232-ac80-7e536c691f94" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.876664] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-9dfe2781-a394-4232-ac80-7e536c691f94" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.069487] env[63197]: DEBUG oslo_vmware.api [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364317, 'name': PowerOnVM_Task, 'duration_secs': 0.421031} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.069764] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.069986] env[63197]: DEBUG nova.compute.manager [None req-e4a89357-6a82-47a8-8bd0-afc5ae9a05c1 tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.071086] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd405f75-d506-4f97-a954-cf1cfaf67ead {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.108547] env[63197]: DEBUG nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.245138] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52783286-fabf-4fea-de21-8ce8a90f569e, 'name': SearchDatastore_Task, 'duration_secs': 0.0142} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.245138] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.245323] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.245547] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.245689] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.246174] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.246174] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-106173b5-1098-4588-9518-182a6a9663e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.254122] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.254122] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 934.254991] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-317bfa81-ebbb-4765-b964-ea19e11f3c69 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.259761] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 934.259761] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528907fc-016e-5ea9-ef55-ce2068e2c411" [ 934.259761] env[63197]: _type = "Task" [ 934.259761] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.267114] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528907fc-016e-5ea9-ef55-ce2068e2c411, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.341876] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "c6686a09-3711-434b-b69f-a2518366fbed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.342164] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "c6686a09-3711-434b-b69f-a2518366fbed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.380446] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.380698] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.381497] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715164af-2261-47ce-86b8-05026a546959 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.402459] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28303960-d04d-4b68-ac3a-ff92c71e4435 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.433195] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfiguring VM to detach interface {{(pid=63197) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 934.433528] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f03acf2-762b-47d0-ac26-8cfc50cec054 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.451619] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 934.451619] env[63197]: value = "task-1364318" [ 934.451619] env[63197]: _type = "Task" [ 934.451619] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.462421] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.537988] env[63197]: DEBUG nova.compute.manager [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Received event network-changed-4bebe341-10f9-4ff8-a0b7-3312eb20074f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.538221] env[63197]: DEBUG nova.compute.manager [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Refreshing instance network info cache due to event network-changed-4bebe341-10f9-4ff8-a0b7-3312eb20074f. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.538440] env[63197]: DEBUG oslo_concurrency.lockutils [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] Acquiring lock "refresh_cache-4757f91b-21f1-4a16-bda3-729b12ddf86c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.538581] env[63197]: DEBUG oslo_concurrency.lockutils [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] Acquired lock "refresh_cache-4757f91b-21f1-4a16-bda3-729b12ddf86c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.538817] env[63197]: DEBUG nova.network.neutron [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Refreshing network info cache for port 4bebe341-10f9-4ff8-a0b7-3312eb20074f {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 934.642272] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.642542] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.643994] env[63197]: INFO nova.compute.claims [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.665654] env[63197]: DEBUG nova.network.neutron [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance_info_cache with network_info: [{"id": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "address": "fa:16:3e:be:d6:04", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d53dd8b-1b", "ovs_interfaceid": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.771347] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528907fc-016e-5ea9-ef55-ce2068e2c411, 'name': SearchDatastore_Task, 'duration_secs': 0.009043} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.772634] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2465d2dc-e64d-4eaf-ad26-05bab735b7d3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.778406] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 934.778406] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280255c-3fff-5afa-5cb3-647b9357feca" [ 934.778406] env[63197]: _type = "Task" [ 934.778406] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.787066] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280255c-3fff-5afa-5cb3-647b9357feca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.845692] env[63197]: DEBUG nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.960813] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.169814] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.255390] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "52549788-b58f-4f5d-8443-a0c771099884" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.255797] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "52549788-b58f-4f5d-8443-a0c771099884" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.256104] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "52549788-b58f-4f5d-8443-a0c771099884-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.256362] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "52549788-b58f-4f5d-8443-a0c771099884-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.256748] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "52549788-b58f-4f5d-8443-a0c771099884-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.262964] env[63197]: INFO nova.compute.manager [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Terminating instance [ 935.264846] env[63197]: DEBUG nova.compute.manager [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 935.265037] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.265853] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8714b1-a9db-43ae-9664-eec5e4a6cdb7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.275466] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 935.275788] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6afb5f25-7482-4681-9333-87944e1a180f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.283669] env[63197]: DEBUG oslo_vmware.api [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 935.283669] env[63197]: value = "task-1364320" [ 935.283669] env[63197]: _type = "Task" [ 935.283669] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.290353] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280255c-3fff-5afa-5cb3-647b9357feca, 'name': SearchDatastore_Task, 'duration_secs': 0.010299} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.290967] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.291244] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 4757f91b-21f1-4a16-bda3-729b12ddf86c/4757f91b-21f1-4a16-bda3-729b12ddf86c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 935.291528] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd0c5b55-4b2c-4ade-99a7-67d404f36de2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.296514] env[63197]: DEBUG oslo_vmware.api [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364320, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.297372] env[63197]: DEBUG nova.network.neutron [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Updated VIF entry in instance network info cache for port 4bebe341-10f9-4ff8-a0b7-3312eb20074f. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 935.297720] env[63197]: DEBUG nova.network.neutron [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Updating instance_info_cache with network_info: [{"id": "4bebe341-10f9-4ff8-a0b7-3312eb20074f", "address": "fa:16:3e:70:fb:da", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4bebe341-10", "ovs_interfaceid": "4bebe341-10f9-4ff8-a0b7-3312eb20074f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.303679] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 935.303679] env[63197]: value = "task-1364321" [ 935.303679] env[63197]: _type = "Task" [ 935.303679] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.312272] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "c97f59db-b1bf-4744-98ed-b6d7c095207f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.312501] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.316844] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.368705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.463402] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.555730] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "638ef9c9-253b-4958-a660-6c1801408a51" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.557936] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.794581] env[63197]: DEBUG oslo_vmware.api [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364320, 'name': PowerOffVM_Task, 'duration_secs': 0.178599} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.794860] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.795108] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.795434] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa52b068-5a53-4779-81be-7e9cb4967c64 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.800409] env[63197]: DEBUG oslo_concurrency.lockutils [req-0654a1fb-f654-429c-a060-3df00e534ca5 req-c93daaf8-2d42-499f-a232-a0ab4a1e3470 service nova] Releasing lock "refresh_cache-4757f91b-21f1-4a16-bda3-729b12ddf86c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.815730] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487547} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.816009] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 4757f91b-21f1-4a16-bda3-729b12ddf86c/4757f91b-21f1-4a16-bda3-729b12ddf86c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 935.816240] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.816582] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34184b9b-cba9-47b1-9f5b-2cef1fcd76d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.818559] env[63197]: DEBUG nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.826131] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 935.826131] env[63197]: value = "task-1364323" [ 935.826131] env[63197]: _type = "Task" [ 935.826131] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.836150] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364323, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.857289] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087f13b1-646e-4b90-a9d8-6991475d4cda {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.860641] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.861118] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.861359] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Deleting the datastore file [datastore1] 52549788-b58f-4f5d-8443-a0c771099884 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.862053] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a1b43ac-b943-4f50-9aba-c5ebc5dbf8b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.867701] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d96c2f2-a154-4569-85c2-0442327f9a86 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.872274] env[63197]: DEBUG oslo_vmware.api [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 935.872274] env[63197]: value = "task-1364324" [ 935.872274] env[63197]: _type = "Task" [ 935.872274] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.902666] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f86f555-4bda-4781-b94a-f632c4d7f785 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.908593] env[63197]: DEBUG oslo_vmware.api [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.913395] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fd361e-3ad2-42cb-9d3d-2251a702e3ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.928572] env[63197]: DEBUG nova.compute.provider_tree [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.962213] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.060020] env[63197]: DEBUG nova.compute.utils [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.337573] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364323, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064803} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.338526] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.338800] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.339660] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99543ed-1821-484c-9196-ae7f9ee7b634 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.360916] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 4757f91b-21f1-4a16-bda3-729b12ddf86c/4757f91b-21f1-4a16-bda3-729b12ddf86c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.361196] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d413e815-573e-417f-b533-89fd781b26f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.383024] env[63197]: DEBUG oslo_vmware.api [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172751} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.384169] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.384356] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 936.384555] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 936.384736] env[63197]: INFO nova.compute.manager [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Took 1.12 seconds to destroy the instance on the hypervisor. [ 936.384965] env[63197]: DEBUG oslo.service.loopingcall [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.385235] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 936.385235] env[63197]: value = "task-1364325" [ 936.385235] env[63197]: _type = "Task" [ 936.385235] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.385421] env[63197]: DEBUG nova.compute.manager [-] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 936.385514] env[63197]: DEBUG nova.network.neutron [-] [instance: 52549788-b58f-4f5d-8443-a0c771099884] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 936.394834] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364325, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.452594] env[63197]: ERROR nova.scheduler.client.report [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [req-08eb7aff-b598-420a-8279-e9d59676fd15] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID e9e01824-bb93-4102-9e5b-66d5dbc1990b. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-08eb7aff-b598-420a-8279-e9d59676fd15"}]} [ 936.464283] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.469179] env[63197]: DEBUG nova.scheduler.client.report [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Refreshing inventories for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 936.487454] env[63197]: DEBUG nova.scheduler.client.report [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Updating ProviderTree inventory for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 936.487454] env[63197]: DEBUG nova.compute.provider_tree [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Updating inventory in ProviderTree for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 936.497178] env[63197]: DEBUG nova.scheduler.client.report [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Refreshing aggregate associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, aggregates: None {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 936.516736] env[63197]: DEBUG nova.scheduler.client.report [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Refreshing trait associations for resource provider e9e01824-bb93-4102-9e5b-66d5dbc1990b, traits: COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,HW_ARCH_X86_64 {{(pid=63197) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 936.563381] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.614460] env[63197]: DEBUG nova.compute.manager [req-e8cce66d-9b3e-40a5-ae56-f7f866b69b05 req-39e97a0e-32db-49de-b232-35bf95a73cc3 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Received event network-vif-deleted-dd634e1d-f13f-4709-a49b-4a669d1286f3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.614686] env[63197]: INFO nova.compute.manager [req-e8cce66d-9b3e-40a5-ae56-f7f866b69b05 req-39e97a0e-32db-49de-b232-35bf95a73cc3 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Neutron deleted interface dd634e1d-f13f-4709-a49b-4a669d1286f3; detaching it from the instance and deleting it from the info cache [ 936.614858] env[63197]: DEBUG nova.network.neutron [req-e8cce66d-9b3e-40a5-ae56-f7f866b69b05 req-39e97a0e-32db-49de-b232-35bf95a73cc3 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.674597] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e2489a1-bef0-4560-a9d7-8cbb047139b2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.681798] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-018ce5f9-d2f0-4f2b-938c-445c11148114 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.688526] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c38f779-80ab-48aa-85ef-5557627e4a72 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.717597] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df56eb2e-9c42-416b-a528-3ba1ff5dcc33 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.732367] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance '537cc8cc-94e0-41b4-8204-6f995f2ea4d3' progress to 0 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 936.740051] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35103e37-c868-4773-bc14-e5161215cbb8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.753270] env[63197]: DEBUG nova.compute.provider_tree [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.896709] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364325, 'name': ReconfigVM_Task, 'duration_secs': 0.261077} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.897034] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 4757f91b-21f1-4a16-bda3-729b12ddf86c/4757f91b-21f1-4a16-bda3-729b12ddf86c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.897627] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a729e68-40f5-4f8e-9f66-e21777afa523 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.903059] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 936.903059] env[63197]: value = "task-1364326" [ 936.903059] env[63197]: _type = "Task" [ 936.903059] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.910084] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364326, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.964271] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.092488] env[63197]: DEBUG nova.network.neutron [-] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.117340] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-467c1cb8-eeab-4be5-b79b-a010c3af3fba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.129504] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795ed481-c9c7-4705-bed8-3afc3fb35c65 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.156201] env[63197]: DEBUG nova.compute.manager [req-e8cce66d-9b3e-40a5-ae56-f7f866b69b05 req-39e97a0e-32db-49de-b232-35bf95a73cc3 service nova] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Detach interface failed, port_id=dd634e1d-f13f-4709-a49b-4a669d1286f3, reason: Instance 52549788-b58f-4f5d-8443-a0c771099884 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 937.196135] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 937.196376] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290360', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'name': 'volume-c29df051-5e11-4357-9d18-1c7dc22c901b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'serial': 'c29df051-5e11-4357-9d18-1c7dc22c901b'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 937.197213] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36775890-9cf1-485e-ac3e-0fc600fb7df6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.213030] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f32c2b-f54c-424a-ad4b-46e06f063f2a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.236179] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] volume-c29df051-5e11-4357-9d18-1c7dc22c901b/volume-c29df051-5e11-4357-9d18-1c7dc22c901b.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.237709] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.237915] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22e36663-86c7-45d6-a599-e3457b8d66f9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.249842] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7969a380-0104-404a-bd0a-68ee1ca11bd4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.256579] env[63197]: DEBUG nova.scheduler.client.report [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.259550] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 937.259550] env[63197]: value = "task-1364328" [ 937.259550] env[63197]: _type = "Task" [ 937.259550] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.261048] env[63197]: DEBUG oslo_vmware.api [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 937.261048] env[63197]: value = "task-1364327" [ 937.261048] env[63197]: _type = "Task" [ 937.261048] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.271494] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364328, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.274713] env[63197]: DEBUG oslo_vmware.api [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364327, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.415135] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364326, 'name': Rename_Task, 'duration_secs': 0.127848} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.415592] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 937.415890] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-923aec68-cc61-4513-b113-fabf1fd770a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.422122] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 937.422122] env[63197]: value = "task-1364329" [ 937.422122] env[63197]: _type = "Task" [ 937.422122] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.432847] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364329, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.465979] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.595476] env[63197]: INFO nova.compute.manager [-] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Took 1.21 seconds to deallocate network for instance. [ 937.624919] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "638ef9c9-253b-4958-a660-6c1801408a51" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.625259] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.625534] env[63197]: INFO nova.compute.manager [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Attaching volume 6665b511-2b56-4af2-8d68-620525b85aa5 to /dev/sdb [ 937.660906] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b440034b-d01b-42d5-8bde-74fb261d61ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.669232] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2160fcbe-0cad-407b-b11b-610153bf2150 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.682314] env[63197]: DEBUG nova.virt.block_device [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updating existing volume attachment record: 6d29ea5e-e68f-484d-82f1-f70d745be274 {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 937.761963] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.119s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.762461] env[63197]: DEBUG nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 937.765182] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.397s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.766580] env[63197]: INFO nova.compute.claims [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.781672] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364328, 'name': PowerOffVM_Task, 'duration_secs': 0.223166} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.784685] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.784853] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance '537cc8cc-94e0-41b4-8204-6f995f2ea4d3' progress to 17 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.789103] env[63197]: DEBUG oslo_vmware.api [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364327, 'name': ReconfigVM_Task, 'duration_secs': 0.326231} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.789816] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfigured VM instance instance-00000044 to attach disk [datastore2] volume-c29df051-5e11-4357-9d18-1c7dc22c901b/volume-c29df051-5e11-4357-9d18-1c7dc22c901b.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.796239] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f4ebac8-6bf8-48c2-9b98-d5cc37678ee8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.810728] env[63197]: DEBUG oslo_vmware.api [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 937.810728] env[63197]: value = "task-1364331" [ 937.810728] env[63197]: _type = "Task" [ 937.810728] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.819019] env[63197]: DEBUG oslo_vmware.api [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364331, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.934562] env[63197]: DEBUG oslo_vmware.api [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364329, 'name': PowerOnVM_Task, 'duration_secs': 0.471568} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.934878] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.935025] env[63197]: INFO nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Took 7.18 seconds to spawn the instance on the hypervisor. [ 937.935268] env[63197]: DEBUG nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.936041] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d08d45-7721-4e41-a1eb-5bd8f86f6da4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.965612] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.102656] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.291994] env[63197]: DEBUG nova.compute.utils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 938.296601] env[63197]: DEBUG nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 938.296875] env[63197]: DEBUG nova.network.neutron [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 938.300399] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.300632] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.300788] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.300972] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.301131] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.301349] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.301476] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.301633] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.301790] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.301949] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.302127] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.308800] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbfc6be2-c453-4d6d-bd14-9d747b2e57b3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.328703] env[63197]: DEBUG oslo_vmware.api [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364331, 'name': ReconfigVM_Task, 'duration_secs': 0.150136} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.330065] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290360', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'name': 'volume-c29df051-5e11-4357-9d18-1c7dc22c901b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'serial': 'c29df051-5e11-4357-9d18-1c7dc22c901b'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 938.332485] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 938.332485] env[63197]: value = "task-1364334" [ 938.332485] env[63197]: _type = "Task" [ 938.332485] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.347178] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364334, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.362512] env[63197]: DEBUG nova.policy [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f2913068578840519ceb7bf29c628116', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4ebe4ad069549aeb83c69abca73e425', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 938.452684] env[63197]: INFO nova.compute.manager [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Took 13.63 seconds to build instance. [ 938.467954] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.627221] env[63197]: DEBUG nova.network.neutron [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Successfully created port: da5d991c-658a-47b4-b74f-ea5cb71fdbc6 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 938.802443] env[63197]: DEBUG nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 938.849172] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364334, 'name': ReconfigVM_Task, 'duration_secs': 0.223871} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.850289] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance '537cc8cc-94e0-41b4-8204-6f995f2ea4d3' progress to 33 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 938.954280] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b989f6b6-ccb0-48c2-9c7c-9431c7376ded tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.140s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.970507] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.014234] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45c4bb2-0cab-4470-8d25-1ab451c92759 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.021229] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6010889-87d5-4452-82ac-d0dbd2c778c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.050449] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88e19da-decc-4aef-9fa3-70ece6e2f485 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.057649] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c35fe8-cee7-435e-a2f7-dbe6f684bbc4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.072030] env[63197]: DEBUG nova.compute.provider_tree [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.361287] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.361541] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.361700] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.361912] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.362129] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.362378] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.362605] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.362764] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.362927] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.363125] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.363303] env[63197]: DEBUG nova.virt.hardware [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.368459] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Reconfiguring VM instance instance-0000004e to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 939.369570] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98bbc3c9-eb9a-4612-8f3a-2ccb245bc38d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.388061] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 939.388061] env[63197]: value = "task-1364335" [ 939.388061] env[63197]: _type = "Task" [ 939.388061] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.396222] env[63197]: DEBUG nova.objects.instance [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lazy-loading 'flavor' on Instance uuid 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.397452] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.467654] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.575211] env[63197]: DEBUG nova.scheduler.client.report [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.648473] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "c41b34cb-cc39-461e-83df-cd4de6780d95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.648705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.814516] env[63197]: DEBUG nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 939.848365] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 939.848618] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 939.848773] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.848948] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 939.849105] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.849264] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 939.849443] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 939.849606] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 939.849841] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 939.850035] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 939.850213] env[63197]: DEBUG nova.virt.hardware [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 939.851106] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4620f00a-fe66-47d7-8ae8-643ee7366e68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.859068] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4259518-f068-47ed-8053-59fa1f4db3a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.896278] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364335, 'name': ReconfigVM_Task, 'duration_secs': 0.258656} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.896535] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Reconfigured VM instance instance-0000004e to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 939.897289] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f16b66-8135-40f8-bdfc-0167a3ef11d9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.901970] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2639625a-ea1a-429c-a6af-be314ad84806 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.353s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.920499] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 537cc8cc-94e0-41b4-8204-6f995f2ea4d3/537cc8cc-94e0-41b4-8204-6f995f2ea4d3.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.921200] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50ebc9c8-70e0-4d3b-b59d-294690c32e41 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.941049] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 939.941049] env[63197]: value = "task-1364336" [ 939.941049] env[63197]: _type = "Task" [ 939.941049] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.948399] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364336, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.966349] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.080385] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.080961] env[63197]: DEBUG nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 940.084864] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.745s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.085428] env[63197]: INFO nova.compute.claims [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.151240] env[63197]: DEBUG nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 940.231638] env[63197]: DEBUG nova.compute.manager [req-2ea1c66c-874c-41a6-af96-9f3f25e4ab44 req-fd360aa2-36d2-4e95-b858-c4d48437d36d service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Received event network-vif-plugged-da5d991c-658a-47b4-b74f-ea5cb71fdbc6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.232488] env[63197]: DEBUG oslo_concurrency.lockutils [req-2ea1c66c-874c-41a6-af96-9f3f25e4ab44 req-fd360aa2-36d2-4e95-b858-c4d48437d36d service nova] Acquiring lock "4d12bac4-da42-42bd-9361-7015f3be9693-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.232687] env[63197]: DEBUG oslo_concurrency.lockutils [req-2ea1c66c-874c-41a6-af96-9f3f25e4ab44 req-fd360aa2-36d2-4e95-b858-c4d48437d36d service nova] Lock "4d12bac4-da42-42bd-9361-7015f3be9693-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.232871] env[63197]: DEBUG oslo_concurrency.lockutils [req-2ea1c66c-874c-41a6-af96-9f3f25e4ab44 req-fd360aa2-36d2-4e95-b858-c4d48437d36d service nova] Lock "4d12bac4-da42-42bd-9361-7015f3be9693-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.233087] env[63197]: DEBUG nova.compute.manager [req-2ea1c66c-874c-41a6-af96-9f3f25e4ab44 req-fd360aa2-36d2-4e95-b858-c4d48437d36d service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] No waiting events found dispatching network-vif-plugged-da5d991c-658a-47b4-b74f-ea5cb71fdbc6 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 940.233323] env[63197]: WARNING nova.compute.manager [req-2ea1c66c-874c-41a6-af96-9f3f25e4ab44 req-fd360aa2-36d2-4e95-b858-c4d48437d36d service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Received unexpected event network-vif-plugged-da5d991c-658a-47b4-b74f-ea5cb71fdbc6 for instance with vm_state building and task_state spawning. [ 940.327713] env[63197]: DEBUG nova.network.neutron [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Successfully updated port: da5d991c-658a-47b4-b74f-ea5cb71fdbc6 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.452029] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364336, 'name': ReconfigVM_Task, 'duration_secs': 0.482772} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.452306] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 537cc8cc-94e0-41b4-8204-6f995f2ea4d3/537cc8cc-94e0-41b4-8204-6f995f2ea4d3.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.452605] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance '537cc8cc-94e0-41b4-8204-6f995f2ea4d3' progress to 50 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.468227] env[63197]: DEBUG oslo_vmware.api [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364318, 'name': ReconfigVM_Task, 'duration_secs': 5.740207} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.468540] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.468805] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Reconfigured VM to detach interface {{(pid=63197) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 940.589972] env[63197]: DEBUG nova.compute.utils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.604181] env[63197]: DEBUG nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.604425] env[63197]: DEBUG nova.network.neutron [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 940.674106] env[63197]: DEBUG nova.policy [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4651bf2d103498cab54fb45a665277b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '61bfcd8038f046789676dc740a9aa991', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.677055] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.692442] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.692697] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.830176] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "refresh_cache-4d12bac4-da42-42bd-9361-7015f3be9693" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.830393] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquired lock "refresh_cache-4d12bac4-da42-42bd-9361-7015f3be9693" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.830459] env[63197]: DEBUG nova.network.neutron [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.934471] env[63197]: DEBUG nova.network.neutron [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Successfully created port: 0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.959504] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1013a61c-fb94-4acd-b5da-9521ea511942 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.980186] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d66f6d-ffba-4191-86cb-66e9812b8cb3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.998063] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance '537cc8cc-94e0-41b4-8204-6f995f2ea4d3' progress to 67 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 941.104054] env[63197]: DEBUG nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 941.196247] env[63197]: DEBUG nova.compute.utils [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 941.295390] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c376f229-95f9-4716-aa1c-7afcabfeecdb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.303588] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53fd4e56-0e00-4268-aa89-fc965e8e60d8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.334429] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91e12a3-4c17-476a-9a7b-52bbcfac69fd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.341495] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cc9ca6-7cda-482a-b4aa-e60419790c14 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.361883] env[63197]: DEBUG nova.compute.provider_tree [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.363930] env[63197]: DEBUG nova.network.neutron [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.492010] env[63197]: DEBUG nova.network.neutron [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Updating instance_info_cache with network_info: [{"id": "da5d991c-658a-47b4-b74f-ea5cb71fdbc6", "address": "fa:16:3e:e0:cc:72", "network": {"id": "36d2fe9a-3846-409b-807f-63e613f7e36f", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-921117937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4ebe4ad069549aeb83c69abca73e425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d991c-65", "ovs_interfaceid": "da5d991c-658a-47b4-b74f-ea5cb71fdbc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.538233] env[63197]: DEBUG nova.network.neutron [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Port 2d53dd8b-1ba1-495d-99fc-265d17cb34a5 binding to destination host cpu-1 is already ACTIVE {{(pid=63197) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 941.699608] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.725232] env[63197]: DEBUG nova.compute.manager [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-vif-deleted-33a215bc-a3d4-48cf-8856-1da55204dbdf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.725232] env[63197]: INFO nova.compute.manager [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Neutron deleted interface 33a215bc-a3d4-48cf-8856-1da55204dbdf; detaching it from the instance and deleting it from the info cache [ 941.725232] env[63197]: DEBUG nova.network.neutron [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9dfe2781-a394-4232-ac80-7e536c691f94", "address": "fa:16:3e:b2:4f:a9", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9dfe2781-a3", "ovs_interfaceid": "9dfe2781-a394-4232-ac80-7e536c691f94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.740349] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.742176] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.742176] env[63197]: DEBUG nova.network.neutron [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.867891] env[63197]: DEBUG nova.scheduler.client.report [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.994608] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Releasing lock "refresh_cache-4d12bac4-da42-42bd-9361-7015f3be9693" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.994981] env[63197]: DEBUG nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Instance network_info: |[{"id": "da5d991c-658a-47b4-b74f-ea5cb71fdbc6", "address": "fa:16:3e:e0:cc:72", "network": {"id": "36d2fe9a-3846-409b-807f-63e613f7e36f", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-921117937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4ebe4ad069549aeb83c69abca73e425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d991c-65", "ovs_interfaceid": "da5d991c-658a-47b4-b74f-ea5cb71fdbc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 941.995437] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:cc:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '42f08482-a1da-405d-9918-d733d9f5173c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da5d991c-658a-47b4-b74f-ea5cb71fdbc6', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.002944] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Creating folder: Project (e4ebe4ad069549aeb83c69abca73e425). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.003207] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ecab294-e49d-412b-9f78-1ab1f7b3d895 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.014664] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Created folder: Project (e4ebe4ad069549aeb83c69abca73e425) in parent group-v290286. [ 942.014884] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Creating folder: Instances. Parent ref: group-v290364. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 942.015127] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4cd34d76-cc5d-451e-a2bb-d8e1185b5039 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.023797] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Created folder: Instances in parent group-v290364. [ 942.024032] env[63197]: DEBUG oslo.service.loopingcall [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.024203] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.024385] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-93b2b0d9-6bca-43a5-8701-fb9debf8e99e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.046050] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.046050] env[63197]: value = "task-1364340" [ 942.046050] env[63197]: _type = "Task" [ 942.046050] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.053517] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364340, 'name': CreateVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.113387] env[63197]: DEBUG nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 942.139196] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.139482] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.139683] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.139922] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.140129] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.140321] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.140562] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.140737] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.140929] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.141132] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.141316] env[63197]: DEBUG nova.virt.hardware [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.142209] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd4b1f69-3b08-497b-a6da-3a9f1b7bfebe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.150314] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423a1c46-33d2-4607-82a6-9d88f6c9191d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.201942] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.202292] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "9c1caa2b-c369-425a-8726-cddadf06f338" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.202593] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.202825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.203038] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "9c1caa2b-c369-425a-8726-cddadf06f338-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.205515] env[63197]: INFO nova.compute.manager [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Terminating instance [ 942.208581] env[63197]: DEBUG nova.compute.manager [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 942.208581] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.208818] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454e0244-1e57-4306-b7a8-e1236c62f6b3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.216315] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.216569] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7d9c028-c658-4eec-9d9d-885183abc36f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.223792] env[63197]: DEBUG oslo_vmware.api [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 942.223792] env[63197]: value = "task-1364341" [ 942.223792] env[63197]: _type = "Task" [ 942.223792] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.228855] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 942.229222] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290363', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'name': 'volume-6665b511-2b56-4af2-8d68-620525b85aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '638ef9c9-253b-4958-a660-6c1801408a51', 'attached_at': '', 'detached_at': '', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'serial': '6665b511-2b56-4af2-8d68-620525b85aa5'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 942.230125] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d11295-4f02-41b3-bbf2-057e828bce29 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.233712] env[63197]: DEBUG oslo_concurrency.lockutils [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] Acquiring lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.237939] env[63197]: DEBUG oslo_vmware.api [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364341, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.252388] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118e3b28-b861-43a6-bb01-fc471290b7b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.277767] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] volume-6665b511-2b56-4af2-8d68-620525b85aa5/volume-6665b511-2b56-4af2-8d68-620525b85aa5.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.279900] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e8a03a1-dcaf-4b80-a5cd-b4bcdbb8f876 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.296215] env[63197]: DEBUG nova.compute.manager [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Received event network-changed-da5d991c-658a-47b4-b74f-ea5cb71fdbc6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.296409] env[63197]: DEBUG nova.compute.manager [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Refreshing instance network info cache due to event network-changed-da5d991c-658a-47b4-b74f-ea5cb71fdbc6. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.296618] env[63197]: DEBUG oslo_concurrency.lockutils [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] Acquiring lock "refresh_cache-4d12bac4-da42-42bd-9361-7015f3be9693" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.296754] env[63197]: DEBUG oslo_concurrency.lockutils [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] Acquired lock "refresh_cache-4d12bac4-da42-42bd-9361-7015f3be9693" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.296906] env[63197]: DEBUG nova.network.neutron [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Refreshing network info cache for port da5d991c-658a-47b4-b74f-ea5cb71fdbc6 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 942.305966] env[63197]: DEBUG oslo_vmware.api [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 942.305966] env[63197]: value = "task-1364342" [ 942.305966] env[63197]: _type = "Task" [ 942.305966] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.314632] env[63197]: DEBUG oslo_vmware.api [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364342, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.373213] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.373804] env[63197]: DEBUG nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.376382] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.274s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.376603] env[63197]: DEBUG nova.objects.instance [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lazy-loading 'resources' on Instance uuid 52549788-b58f-4f5d-8443-a0c771099884 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.514021] env[63197]: DEBUG nova.network.neutron [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Successfully updated port: 0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.554270] env[63197]: INFO nova.network.neutron [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Port 9dfe2781-a394-4232-ac80-7e536c691f94 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 942.554595] env[63197]: DEBUG nova.network.neutron [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.566074] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.566355] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.566544] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.575127] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364340, 'name': CreateVM_Task, 'duration_secs': 0.273514} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.576278] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.576873] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.577055] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.577405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 942.577890] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39fa6d67-1f1b-40bb-b21d-395df94de63e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.583174] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 942.583174] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5260ff7b-7fe6-ccd2-41cf-af97d68a6e28" [ 942.583174] env[63197]: _type = "Task" [ 942.583174] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.320010] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.320325] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.320545] env[63197]: INFO nova.compute.manager [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Attaching volume 07d531bd-d1e0-40e3-8f73-caa2955b6150 to /dev/sdc [ 943.325054] env[63197]: DEBUG nova.compute.utils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.328364] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.328496] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.328635] env[63197]: DEBUG nova.network.neutron [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.329723] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.340954] env[63197]: DEBUG nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.341119] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 943.352523] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5260ff7b-7fe6-ccd2-41cf-af97d68a6e28, 'name': SearchDatastore_Task, 'duration_secs': 0.009259} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.353112] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.353362] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 943.353607] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.353751] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.353945] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 943.361457] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09a7d67d-c446-42a8-b0d7-2399867fc13a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.366526] env[63197]: DEBUG oslo_vmware.api [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364342, 'name': ReconfigVM_Task, 'duration_secs': 0.381295} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.366759] env[63197]: DEBUG oslo_vmware.api [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364341, 'name': PowerOffVM_Task, 'duration_secs': 0.240855} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.367306] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Reconfigured VM instance instance-00000047 to attach disk [datastore1] volume-6665b511-2b56-4af2-8d68-620525b85aa5/volume-6665b511-2b56-4af2-8d68-620525b85aa5.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.372360] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.372537] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 943.376269] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e06dbdc6-fbd5-4dd8-89d7-aea89b82f4e2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.386767] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe29d5b2-3ca7-4329-867e-d2d9960420e9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.392314] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 943.392314] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 943.394031] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-048aee9d-460b-47be-b351-48b74d5dc456 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.397190] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51741ec8-2051-4a7f-b737-e436eb383c9b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.404787] env[63197]: DEBUG oslo_vmware.api [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 943.404787] env[63197]: value = "task-1364344" [ 943.404787] env[63197]: _type = "Task" [ 943.404787] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.415666] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 943.415666] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5254653c-74eb-b00a-cf1a-0a4ea2c6acde" [ 943.415666] env[63197]: _type = "Task" [ 943.415666] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.425032] env[63197]: DEBUG oslo_vmware.api [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364344, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.425939] env[63197]: DEBUG nova.policy [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91bab7add6374ce4adce53b2595dc5a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4f082d2f82c42218116efa35aee1f56', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 943.427967] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec603605-5ed4-4e08-8392-72bd0252775e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.437087] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5254653c-74eb-b00a-cf1a-0a4ea2c6acde, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.438702] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67c202ed-19a4-427d-be1e-a9bd4f135949 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.444961] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 943.444961] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52975045-59d1-c78d-12a1-d675d3dbaceb" [ 943.444961] env[63197]: _type = "Task" [ 943.444961] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.452233] env[63197]: DEBUG nova.virt.block_device [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updating existing volume attachment record: 16dcbe8d-e51b-4b15-993f-008ea77d2771 {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 943.463911] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52975045-59d1-c78d-12a1-d675d3dbaceb, 'name': SearchDatastore_Task, 'duration_secs': 0.011284} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.464226] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.464545] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 4d12bac4-da42-42bd-9361-7015f3be9693/4d12bac4-da42-42bd-9361-7015f3be9693.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 943.464850] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbda2fc5-654f-4288-939d-ec9f2d584e4c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.474146] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 943.474146] env[63197]: value = "task-1364345" [ 943.474146] env[63197]: _type = "Task" [ 943.474146] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.482675] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364345, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.521854] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 943.522061] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 943.522324] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleting the datastore file [datastore1] 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.522991] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db40f126-957b-4dff-abf4-182d00bc0147 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.530573] env[63197]: DEBUG oslo_vmware.api [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 943.530573] env[63197]: value = "task-1364346" [ 943.530573] env[63197]: _type = "Task" [ 943.530573] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.541738] env[63197]: DEBUG oslo_vmware.api [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.631590] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40daf7e4-b867-400a-8e6a-0b3afd4f817c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.639941] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b1f16c-454e-448b-a652-9ad9c66b1a34 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.675785] env[63197]: DEBUG nova.network.neutron [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Updated VIF entry in instance network info cache for port da5d991c-658a-47b4-b74f-ea5cb71fdbc6. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 943.676176] env[63197]: DEBUG nova.network.neutron [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Updating instance_info_cache with network_info: [{"id": "da5d991c-658a-47b4-b74f-ea5cb71fdbc6", "address": "fa:16:3e:e0:cc:72", "network": {"id": "36d2fe9a-3846-409b-807f-63e613f7e36f", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-921117937-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4ebe4ad069549aeb83c69abca73e425", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "42f08482-a1da-405d-9918-d733d9f5173c", "external-id": "nsx-vlan-transportzone-381", "segmentation_id": 381, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5d991c-65", "ovs_interfaceid": "da5d991c-658a-47b4-b74f-ea5cb71fdbc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.678058] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6babf159-9ce6-4aeb-815a-79f1af03f523 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.686663] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc27273-9bc6-4210-af1a-9b299ebc03bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.702468] env[63197]: DEBUG nova.compute.provider_tree [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.801902] env[63197]: DEBUG nova.compute.manager [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Received event network-vif-plugged-0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.802379] env[63197]: DEBUG oslo_concurrency.lockutils [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] Acquiring lock "c6686a09-3711-434b-b69f-a2518366fbed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.802636] env[63197]: DEBUG oslo_concurrency.lockutils [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] Lock "c6686a09-3711-434b-b69f-a2518366fbed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.802869] env[63197]: DEBUG oslo_concurrency.lockutils [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] Lock "c6686a09-3711-434b-b69f-a2518366fbed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.803108] env[63197]: DEBUG nova.compute.manager [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] No waiting events found dispatching network-vif-plugged-0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.803164] env[63197]: WARNING nova.compute.manager [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Received unexpected event network-vif-plugged-0bdd3cb1-b50c-4791-9d45-3a1766a307da for instance with vm_state building and task_state spawning. [ 943.803304] env[63197]: DEBUG nova.compute.manager [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Received event network-changed-0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.803469] env[63197]: DEBUG nova.compute.manager [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Refreshing instance network info cache due to event network-changed-0bdd3cb1-b50c-4791-9d45-3a1766a307da. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.803722] env[63197]: DEBUG oslo_concurrency.lockutils [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] Acquiring lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.837774] env[63197]: DEBUG nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.841353] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4988aea8-592b-4929-8fb0-996a2aa946ce tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-9c1caa2b-c369-425a-8726-cddadf06f338-9dfe2781-a394-4232-ac80-7e536c691f94" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.966s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.846827] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Successfully created port: c31363ef-bdd6-49f5-89f7-fb36e583c894 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.867760] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.867949] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.868165] env[63197]: INFO nova.compute.manager [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Shelving [ 943.883336] env[63197]: DEBUG nova.network.neutron [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.922036] env[63197]: DEBUG oslo_vmware.api [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364344, 'name': ReconfigVM_Task, 'duration_secs': 0.175186} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.922459] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290363', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'name': 'volume-6665b511-2b56-4af2-8d68-620525b85aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '638ef9c9-253b-4958-a660-6c1801408a51', 'attached_at': '', 'detached_at': '', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'serial': '6665b511-2b56-4af2-8d68-620525b85aa5'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 943.981879] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.984035] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.984035] env[63197]: DEBUG nova.network.neutron [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.991866] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364345, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.043157] env[63197]: DEBUG oslo_vmware.api [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.475998} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.043578] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.043884] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 944.044184] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 944.044481] env[63197]: INFO nova.compute.manager [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Took 1.84 seconds to destroy the instance on the hypervisor. [ 944.044851] env[63197]: DEBUG oslo.service.loopingcall [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.045152] env[63197]: DEBUG nova.compute.manager [-] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 944.045343] env[63197]: DEBUG nova.network.neutron [-] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 944.164857] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Successfully created port: 720e1317-71ba-4079-a3fb-794d854e0e13 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 944.182388] env[63197]: DEBUG oslo_concurrency.lockutils [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] Releasing lock "refresh_cache-4d12bac4-da42-42bd-9361-7015f3be9693" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.182828] env[63197]: DEBUG nova.compute.manager [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-vif-deleted-9dfe2781-a394-4232-ac80-7e536c691f94 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.183146] env[63197]: INFO nova.compute.manager [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Neutron deleted interface 9dfe2781-a394-4232-ac80-7e536c691f94; detaching it from the instance and deleting it from the info cache [ 944.183523] env[63197]: DEBUG nova.network.neutron [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [{"id": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "address": "fa:16:3e:1a:3f:50", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d3f4b96-07", "ovs_interfaceid": "1d3f4b96-07ba-4875-9441-753c3b9855a9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.209023] env[63197]: DEBUG nova.scheduler.client.report [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.295241] env[63197]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 33a215bc-a3d4-48cf-8856-1da55204dbdf could not be found.", "detail": ""}} {{(pid=63197) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 944.295241] env[63197]: DEBUG nova.network.neutron [-] Unable to show port 33a215bc-a3d4-48cf-8856-1da55204dbdf as it no longer exists. {{(pid=63197) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 944.378877] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.378877] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7f65923-872e-40a3-8d85-681f35026e02 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.386047] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 944.386047] env[63197]: value = "task-1364348" [ 944.386047] env[63197]: _type = "Task" [ 944.386047] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.389921] env[63197]: DEBUG nova.network.neutron [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updating instance_info_cache with network_info: [{"id": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "address": "fa:16:3e:49:a5:9e", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bdd3cb1-b5", "ovs_interfaceid": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.397913] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364348, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.487098] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364345, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539743} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.489318] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 4d12bac4-da42-42bd-9361-7015f3be9693/4d12bac4-da42-42bd-9361-7015f3be9693.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.489562] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.490221] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bead1b5e-dfce-4074-bc41-7a9077533110 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.497407] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 944.497407] env[63197]: value = "task-1364349" [ 944.497407] env[63197]: _type = "Task" [ 944.497407] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.505278] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.688633] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d15c2da4-2b52-4ba3-befe-e808ac0e15b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.698518] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8592f40b-72a1-4499-a53a-591698663cf0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.712118] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.335s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.714838] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.038s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.716487] env[63197]: INFO nova.compute.claims [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.737308] env[63197]: DEBUG nova.compute.manager [req-89eeeeb7-85b7-4bdf-acdb-2f5aa602300c req-772b46aa-b739-4615-9c4c-25397f99c025 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Detach interface failed, port_id=9dfe2781-a394-4232-ac80-7e536c691f94, reason: Instance 9c1caa2b-c369-425a-8726-cddadf06f338 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 944.744087] env[63197]: INFO nova.scheduler.client.report [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Deleted allocations for instance 52549788-b58f-4f5d-8443-a0c771099884 [ 944.849566] env[63197]: DEBUG nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.876844] env[63197]: DEBUG nova.network.neutron [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance_info_cache with network_info: [{"id": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "address": "fa:16:3e:be:d6:04", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d53dd8b-1b", "ovs_interfaceid": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.882714] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.882714] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.882879] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.882966] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.883121] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.883348] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.883568] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.883725] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.883884] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.884055] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.884342] env[63197]: DEBUG nova.virt.hardware [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.886323] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ae42d1-dad3-4795-a735-cec42487de1c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.893418] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Releasing lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.893694] env[63197]: DEBUG nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Instance network_info: |[{"id": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "address": "fa:16:3e:49:a5:9e", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bdd3cb1-b5", "ovs_interfaceid": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 944.896782] env[63197]: DEBUG oslo_concurrency.lockutils [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] Acquired lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.896957] env[63197]: DEBUG nova.network.neutron [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Refreshing network info cache for port 0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.898287] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:a5:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e30245c5-78f5-48e6-b504-c6c21f5a9b45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0bdd3cb1-b50c-4791-9d45-3a1766a307da', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.907295] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Creating folder: Project (61bfcd8038f046789676dc740a9aa991). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 944.907734] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364348, 'name': PowerOffVM_Task, 'duration_secs': 0.180574} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.913026] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8927c200-9837-4ae8-876b-a83eb98dc586 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.913835] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 944.917220] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5aeaf0e-8b24-41ea-ba3e-f7cb673422f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.921769] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979103d9-8a51-4bb7-8513-e25d71b50af1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.951486] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Created folder: Project (61bfcd8038f046789676dc740a9aa991) in parent group-v290286. [ 944.951708] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Creating folder: Instances. Parent ref: group-v290368. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 944.952659] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-64a17b40-a87d-4697-8fc3-4ec0b1a2841d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.954601] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e222454e-8a7a-4192-a45c-6d17100cef5e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.967202] env[63197]: DEBUG nova.objects.instance [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'flavor' on Instance uuid 638ef9c9-253b-4958-a660-6c1801408a51 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.967202] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Created folder: Instances in parent group-v290368. [ 944.967202] env[63197]: DEBUG oslo.service.loopingcall [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.967887] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 944.967887] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-344052e7-81f2-4e92-bf9d-6b40718f0e8e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.988133] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.988133] env[63197]: value = "task-1364352" [ 944.988133] env[63197]: _type = "Task" [ 944.988133] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.995251] env[63197]: DEBUG nova.compute.manager [req-701445d8-0eef-4223-836e-479a2ceca9dc req-c19d56c4-cac4-453d-9bbe-0c63d19c0707 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Received event network-vif-deleted-1d3f4b96-07ba-4875-9441-753c3b9855a9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.995568] env[63197]: INFO nova.compute.manager [req-701445d8-0eef-4223-836e-479a2ceca9dc req-c19d56c4-cac4-453d-9bbe-0c63d19c0707 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Neutron deleted interface 1d3f4b96-07ba-4875-9441-753c3b9855a9; detaching it from the instance and deleting it from the info cache [ 944.995633] env[63197]: DEBUG nova.network.neutron [req-701445d8-0eef-4223-836e-479a2ceca9dc req-c19d56c4-cac4-453d-9bbe-0c63d19c0707 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.005230] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364352, 'name': CreateVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.010244] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06755} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.010485] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.011225] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aff0c67-8449-4ba7-8c30-dd8dcc33cb09 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.035974] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 4d12bac4-da42-42bd-9361-7015f3be9693/4d12bac4-da42-42bd-9361-7015f3be9693.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.036307] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-783fb794-a552-4db9-980a-09476225b406 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.056843] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 945.056843] env[63197]: value = "task-1364353" [ 945.056843] env[63197]: _type = "Task" [ 945.056843] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.065326] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364353, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.250980] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1fa2d086-8208-4425-928c-df90f4721c2f tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "52549788-b58f-4f5d-8443-a0c771099884" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.995s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.379100] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.417237] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.419906] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.419906] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.419906] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.419906] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.420606] env[63197]: INFO nova.compute.manager [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Terminating instance [ 945.422790] env[63197]: DEBUG nova.compute.manager [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.426017] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.426017] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042ed8a4-65e1-48bf-a7d4-2e5a51b7e2e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.434897] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.435177] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a6c7bde-d520-4390-b465-6f8b0868afbc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.440829] env[63197]: DEBUG oslo_vmware.api [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 945.440829] env[63197]: value = "task-1364354" [ 945.440829] env[63197]: _type = "Task" [ 945.440829] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.452339] env[63197]: DEBUG oslo_vmware.api [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364354, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.466269] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 945.468716] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7a3e9e87-2515-4a16-b3d6-22ccb05c3605 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.471300] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1812ae74-d5bc-4a3c-9b3d-d55d13c3fb51 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.846s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.481035] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 945.481035] env[63197]: value = "task-1364355" [ 945.481035] env[63197]: _type = "Task" [ 945.481035] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.483943] env[63197]: DEBUG nova.network.neutron [-] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.491884] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364355, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.499625] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42a02184-685a-461c-9dd2-dbd368926fae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.505910] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364352, 'name': CreateVM_Task, 'duration_secs': 0.358752} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.506548] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.507639] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.507811] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.508210] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.508481] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19859e15-d32c-414e-9cea-ab6e22828642 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.515540] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a84a7f5-f88c-4950-9c07-6491e387480b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.529929] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 945.529929] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5283ae10-552b-f22e-596d-97893042325a" [ 945.529929] env[63197]: _type = "Task" [ 945.529929] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.548421] env[63197]: DEBUG nova.compute.manager [req-701445d8-0eef-4223-836e-479a2ceca9dc req-c19d56c4-cac4-453d-9bbe-0c63d19c0707 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Detach interface failed, port_id=1d3f4b96-07ba-4875-9441-753c3b9855a9, reason: Instance 9c1caa2b-c369-425a-8726-cddadf06f338 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 945.552392] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5283ae10-552b-f22e-596d-97893042325a, 'name': SearchDatastore_Task, 'duration_secs': 0.009766} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.554619] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.554619] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.554619] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.554619] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.554619] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.554619] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-90885c06-1cb0-42f3-bdf8-b6ac38c185b1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.564221] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.564221] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.564382] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c01800b-5207-4af5-a2a4-43c9bf6796f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.569509] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364353, 'name': ReconfigVM_Task, 'duration_secs': 0.256188} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.569981] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 4d12bac4-da42-42bd-9361-7015f3be9693/4d12bac4-da42-42bd-9361-7015f3be9693.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.570656] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caccf307-5c8e-43b1-8c31-0116969375be {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.574024] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 945.574024] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f62b5f-c098-7b89-7a17-405ad092a6c7" [ 945.574024] env[63197]: _type = "Task" [ 945.574024] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.579086] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 945.579086] env[63197]: value = "task-1364356" [ 945.579086] env[63197]: _type = "Task" [ 945.579086] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.582080] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f62b5f-c098-7b89-7a17-405ad092a6c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.590843] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364356, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.683930] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Successfully updated port: c31363ef-bdd6-49f5-89f7-fb36e583c894 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.784909] env[63197]: DEBUG nova.network.neutron [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updated VIF entry in instance network info cache for port 0bdd3cb1-b50c-4791-9d45-3a1766a307da. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.785334] env[63197]: DEBUG nova.network.neutron [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updating instance_info_cache with network_info: [{"id": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "address": "fa:16:3e:49:a5:9e", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bdd3cb1-b5", "ovs_interfaceid": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.910400] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1631aab3-dd82-4a44-a8bd-721021819e1d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.931228] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3f24cb-6e45-4eec-afe1-83b9aca5f474 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.938900] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance '537cc8cc-94e0-41b4-8204-6f995f2ea4d3' progress to 83 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 945.959652] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec70a8f1-d8df-4f44-bf27-2cc8e43f23e1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.962431] env[63197]: DEBUG oslo_vmware.api [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364354, 'name': PowerOffVM_Task, 'duration_secs': 0.289909} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.962837] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.963147] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.963892] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d31a8d55-0994-44bf-865f-8ff4ee6cc156 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.968618] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40d37761-94c6-40c6-8fca-c35071462ab1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.002709] env[63197]: INFO nova.compute.manager [-] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Took 1.96 seconds to deallocate network for instance. [ 946.009159] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ddb8e3-a917-406e-afaa-9ac43a36dccb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.019394] env[63197]: DEBUG nova.compute.manager [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received event network-vif-plugged-c31363ef-bdd6-49f5-89f7-fb36e583c894 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.019394] env[63197]: DEBUG oslo_concurrency.lockutils [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] Acquiring lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.019394] env[63197]: DEBUG oslo_concurrency.lockutils [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.019394] env[63197]: DEBUG oslo_concurrency.lockutils [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.019394] env[63197]: DEBUG nova.compute.manager [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] No waiting events found dispatching network-vif-plugged-c31363ef-bdd6-49f5-89f7-fb36e583c894 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.019394] env[63197]: WARNING nova.compute.manager [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received unexpected event network-vif-plugged-c31363ef-bdd6-49f5-89f7-fb36e583c894 for instance with vm_state building and task_state spawning. [ 946.019394] env[63197]: DEBUG nova.compute.manager [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received event network-changed-c31363ef-bdd6-49f5-89f7-fb36e583c894 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.019394] env[63197]: DEBUG nova.compute.manager [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Refreshing instance network info cache due to event network-changed-c31363ef-bdd6-49f5-89f7-fb36e583c894. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.019394] env[63197]: DEBUG oslo_concurrency.lockutils [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] Acquiring lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.019394] env[63197]: DEBUG oslo_concurrency.lockutils [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] Acquired lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.019899] env[63197]: DEBUG nova.network.neutron [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Refreshing network info cache for port c31363ef-bdd6-49f5-89f7-fb36e583c894 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.026138] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364355, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.027497] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0af222-aa08-4648-9855-3b03680fda62 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.042862] env[63197]: DEBUG nova.compute.provider_tree [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.053348] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.053588] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.053790] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Deleting the datastore file [datastore2] 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.054016] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4bd982d7-9f7c-47a9-8e3e-7c02bd6a166a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.061491] env[63197]: DEBUG oslo_vmware.api [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for the task: (returnval){ [ 946.061491] env[63197]: value = "task-1364359" [ 946.061491] env[63197]: _type = "Task" [ 946.061491] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.072040] env[63197]: DEBUG oslo_vmware.api [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364359, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.083930] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f62b5f-c098-7b89-7a17-405ad092a6c7, 'name': SearchDatastore_Task, 'duration_secs': 0.00995} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.090025] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5574e74c-cba3-430f-9fec-b542c1d3a650 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.094204] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364356, 'name': Rename_Task, 'duration_secs': 0.132813} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.095413] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.095718] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 946.095718] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aaba07-c42e-7fdc-6e45-d353a8d2c1fb" [ 946.095718] env[63197]: _type = "Task" [ 946.095718] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.095902] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dee45943-82b6-4e08-8783-f2df4f3a12df {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.106016] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aaba07-c42e-7fdc-6e45-d353a8d2c1fb, 'name': SearchDatastore_Task, 'duration_secs': 0.009521} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.107154] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.107411] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] c6686a09-3711-434b-b69f-a2518366fbed/c6686a09-3711-434b-b69f-a2518366fbed.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.107702] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 946.107702] env[63197]: value = "task-1364360" [ 946.107702] env[63197]: _type = "Task" [ 946.107702] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.107878] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f3c118d-745c-4c1e-8e4d-4818c5ba10c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.117896] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.119209] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 946.119209] env[63197]: value = "task-1364361" [ 946.119209] env[63197]: _type = "Task" [ 946.119209] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.126372] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364361, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.291039] env[63197]: DEBUG oslo_concurrency.lockutils [req-4e4013e7-3b2d-49e3-9530-1aba2893ecaa req-b784e8c5-bcd6-44c1-8a72-fb84e45bf419 service nova] Releasing lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.449879] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 946.450316] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a50926e6-d0c6-4007-9298-e757e9fab4f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.461077] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 946.461077] env[63197]: value = "task-1364362" [ 946.461077] env[63197]: _type = "Task" [ 946.461077] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.473897] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.514902] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364355, 'name': CreateSnapshot_Task, 'duration_secs': 0.595807} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.515247] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 946.516262] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34cb07d3-2319-405f-9f64-e4463a0ac6a1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.526765] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.546051] env[63197]: DEBUG nova.scheduler.client.report [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.576957] env[63197]: DEBUG oslo_vmware.api [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Task: {'id': task-1364359, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204731} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.578092] env[63197]: DEBUG nova.network.neutron [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.580367] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.582496] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.582496] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.582496] env[63197]: INFO nova.compute.manager [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 946.582496] env[63197]: DEBUG oslo.service.loopingcall [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.582496] env[63197]: DEBUG nova.compute.manager [-] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.582496] env[63197]: DEBUG nova.network.neutron [-] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 946.620616] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364360, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.629839] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364361, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.681498] env[63197]: DEBUG nova.network.neutron [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.971514] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364362, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.039458] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 947.039844] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c73206ac-3b9c-4de6-8416-a9dc8b9cc6ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.049023] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 947.049023] env[63197]: value = "task-1364363" [ 947.049023] env[63197]: _type = "Task" [ 947.049023] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.051318] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.051797] env[63197]: DEBUG nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.054995] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.528s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.055227] env[63197]: DEBUG nova.objects.instance [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'resources' on Instance uuid 9c1caa2b-c369-425a-8726-cddadf06f338 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.059895] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364363, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.121483] env[63197]: DEBUG oslo_vmware.api [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364360, 'name': PowerOnVM_Task, 'duration_secs': 0.628331} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.125177] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.125440] env[63197]: INFO nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Took 7.31 seconds to spawn the instance on the hypervisor. [ 947.125651] env[63197]: DEBUG nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.126494] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84868127-e281-45b9-9bd0-37e2feb70a6e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.138935] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364361, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512756} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.138935] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] c6686a09-3711-434b-b69f-a2518366fbed/c6686a09-3711-434b-b69f-a2518366fbed.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.139197] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.139425] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f9058ee-f4dc-4c6a-b1a3-28fac47382e4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.146309] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 947.146309] env[63197]: value = "task-1364364" [ 947.146309] env[63197]: _type = "Task" [ 947.146309] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.156747] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364364, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.181179] env[63197]: DEBUG nova.compute.manager [req-af654c46-02b4-49b9-a053-3546ffdbf3d6 req-6b0f5fa0-2a50-4896-b09b-0a07aa6adc62 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Received event network-vif-deleted-a42d56b7-bb6b-40d8-a491-2304a1e4a65b {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.181275] env[63197]: INFO nova.compute.manager [req-af654c46-02b4-49b9-a053-3546ffdbf3d6 req-6b0f5fa0-2a50-4896-b09b-0a07aa6adc62 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Neutron deleted interface a42d56b7-bb6b-40d8-a491-2304a1e4a65b; detaching it from the instance and deleting it from the info cache [ 947.181445] env[63197]: DEBUG nova.network.neutron [req-af654c46-02b4-49b9-a053-3546ffdbf3d6 req-6b0f5fa0-2a50-4896-b09b-0a07aa6adc62 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.184453] env[63197]: DEBUG oslo_concurrency.lockutils [req-e812880c-0d02-4123-897d-6f29632ad2e5 req-e21cd69c-14d4-431f-9151-57c68928d7b3 service nova] Releasing lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.336238] env[63197]: DEBUG nova.network.neutron [-] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.473117] env[63197]: DEBUG oslo_vmware.api [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364362, 'name': PowerOnVM_Task, 'duration_secs': 0.666709} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.473683] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 947.473683] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1d4733-3973-4881-a5f5-b2147c902791 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance '537cc8cc-94e0-41b4-8204-6f995f2ea4d3' progress to 100 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 947.558337] env[63197]: DEBUG nova.compute.utils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 947.561887] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364363, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.563573] env[63197]: DEBUG nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 947.563573] env[63197]: DEBUG nova.network.neutron [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.656791] env[63197]: INFO nova.compute.manager [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Took 13.04 seconds to build instance. [ 947.660609] env[63197]: DEBUG nova.policy [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12d00044b65c48f484ac9fea595d6179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80138785cfdb4a2188ffb972da437646', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 947.669125] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364364, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.233109} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.669393] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.670700] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43929e8e-30ed-4a00-9017-0fc948cde1af {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.692886] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] c6686a09-3711-434b-b69f-a2518366fbed/c6686a09-3711-434b-b69f-a2518366fbed.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.695746] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d6ae4627-53f9-42ac-ad6c-051afe631efa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.699130] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e4266bd-d3af-4f94-8af3-fb3218fa7d3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.724295] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627c3413-e501-4194-80df-3bd839c9ee46 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.736757] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 947.736757] env[63197]: value = "task-1364365" [ 947.736757] env[63197]: _type = "Task" [ 947.736757] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.746675] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364365, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.756425] env[63197]: DEBUG nova.compute.manager [req-af654c46-02b4-49b9-a053-3546ffdbf3d6 req-6b0f5fa0-2a50-4896-b09b-0a07aa6adc62 service nova] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Detach interface failed, port_id=a42d56b7-bb6b-40d8-a491-2304a1e4a65b, reason: Instance 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 947.806711] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf148e7-a6bf-4993-b70c-f08eda15cb2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.816187] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "8e5ef1b0-7532-498a-84c2-189274a36c50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.816470] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.823629] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-171ca0c0-b5ca-4c34-93d3-45eb76f2e929 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.857230] env[63197]: INFO nova.compute.manager [-] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Took 1.27 seconds to deallocate network for instance. [ 947.860317] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf0b38ff-af88-4e42-8c23-abd54664593c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.873812] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab788e3-f691-4ef8-8c6e-9acc2c62ffd0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.887612] env[63197]: DEBUG nova.compute.provider_tree [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.036331] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 948.036711] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290367', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'name': 'volume-07d531bd-d1e0-40e3-8f73-caa2955b6150', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'serial': '07d531bd-d1e0-40e3-8f73-caa2955b6150'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 948.037840] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45db35ea-1e81-44ec-bc0e-5a5145d79dc7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.067021] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1093c9-96b7-463e-9d25-90f2b408bccc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.071192] env[63197]: DEBUG nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.072109] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Successfully updated port: 720e1317-71ba-4079-a3fb-794d854e0e13 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.095375] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364363, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.104240] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] volume-07d531bd-d1e0-40e3-8f73-caa2955b6150/volume-07d531bd-d1e0-40e3-8f73-caa2955b6150.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.104834] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04b1ef4e-0422-483a-b402-a45d41b17848 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.121246] env[63197]: DEBUG nova.network.neutron [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Successfully created port: e33a7c35-f308-47b5-8b23-08328d08d9e5 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.130020] env[63197]: DEBUG oslo_vmware.api [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 948.130020] env[63197]: value = "task-1364366" [ 948.130020] env[63197]: _type = "Task" [ 948.130020] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.139629] env[63197]: DEBUG oslo_vmware.api [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.159988] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8669f3a9-05d2-4f50-a78b-d547b14fdf24 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "4d12bac4-da42-42bd-9361-7015f3be9693" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.554s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.247042] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364365, 'name': ReconfigVM_Task, 'duration_secs': 0.26719} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.247362] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Reconfigured VM instance instance-00000051 to attach disk [datastore1] c6686a09-3711-434b-b69f-a2518366fbed/c6686a09-3711-434b-b69f-a2518366fbed.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.248118] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f63a7e7b-d94c-4166-b2c6-768d078453f5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.254291] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 948.254291] env[63197]: value = "task-1364367" [ 948.254291] env[63197]: _type = "Task" [ 948.254291] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.265479] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364367, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.323210] env[63197]: DEBUG nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 948.372585] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.392083] env[63197]: DEBUG nova.scheduler.client.report [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.570032] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364363, 'name': CloneVM_Task, 'duration_secs': 1.453135} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.571025] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Created linked-clone VM from snapshot [ 948.571168] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1701dd97-42b8-4e98-896a-3224943b5e44 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.578833] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Uploading image 073c5ea5-cd55-4c9a-ab31-9918df041ed1 {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 948.601028] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 948.601028] env[63197]: value = "vm-290372" [ 948.601028] env[63197]: _type = "VirtualMachine" [ 948.601028] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 948.601290] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bd6fdf60-16e8-4fac-9e58-faa4f9f0a4a3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.604154] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.604154] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.604154] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.608888] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease: (returnval){ [ 948.608888] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f9b54-45be-2c7b-439d-f46620253045" [ 948.608888] env[63197]: _type = "HttpNfcLease" [ 948.608888] env[63197]: } obtained for exporting VM: (result){ [ 948.608888] env[63197]: value = "vm-290372" [ 948.608888] env[63197]: _type = "VirtualMachine" [ 948.608888] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 948.609154] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the lease: (returnval){ [ 948.609154] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f9b54-45be-2c7b-439d-f46620253045" [ 948.609154] env[63197]: _type = "HttpNfcLease" [ 948.609154] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 948.616739] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 948.616739] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f9b54-45be-2c7b-439d-f46620253045" [ 948.616739] env[63197]: _type = "HttpNfcLease" [ 948.616739] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 948.640922] env[63197]: DEBUG oslo_vmware.api [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364366, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.764475] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364367, 'name': Rename_Task, 'duration_secs': 0.133724} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.764739] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.765035] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3ba9a28a-a90e-44f0-abf0-addc82c57650 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.771332] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 948.771332] env[63197]: value = "task-1364369" [ 948.771332] env[63197]: _type = "Task" [ 948.771332] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.780026] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.852688] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.896733] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.899102] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.527s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.899349] env[63197]: DEBUG nova.objects.instance [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lazy-loading 'resources' on Instance uuid 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.918414] env[63197]: INFO nova.scheduler.client.report [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleted allocations for instance 9c1caa2b-c369-425a-8726-cddadf06f338 [ 949.104909] env[63197]: DEBUG nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.117683] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 949.117683] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f9b54-45be-2c7b-439d-f46620253045" [ 949.117683] env[63197]: _type = "HttpNfcLease" [ 949.117683] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 949.117976] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 949.117976] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f9b54-45be-2c7b-439d-f46620253045" [ 949.117976] env[63197]: _type = "HttpNfcLease" [ 949.117976] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 949.118719] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdeccb0-2711-4ab5-9517-c905a53ea14f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.126577] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ebcd2-f6e3-f04a-e49d-281a1a6e1d2d/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 949.126775] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ebcd2-f6e3-f04a-e49d-281a1a6e1d2d/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 949.130240] env[63197]: DEBUG nova.compute.manager [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.131483] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c2cb80-c0e7-4e21-afe8-642d951aea3b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.194046] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.194309] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.194533] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.194681] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.194842] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.194988] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.195245] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.195469] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.195657] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.195846] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.196118] env[63197]: DEBUG nova.virt.hardware [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.198617] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6bf140-0695-4c76-8424-5e173c088f5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.202402] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 949.214688] env[63197]: DEBUG oslo_vmware.api [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364366, 'name': ReconfigVM_Task, 'duration_secs': 0.706807} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.215326] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfigured VM instance instance-00000044 to attach disk [datastore1] volume-07d531bd-d1e0-40e3-8f73-caa2955b6150/volume-07d531bd-d1e0-40e3-8f73-caa2955b6150.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.223938] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-848b80c6-d253-4019-8ade-3ebd3d136ca6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.235091] env[63197]: DEBUG nova.compute.manager [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received event network-vif-plugged-720e1317-71ba-4079-a3fb-794d854e0e13 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.235297] env[63197]: DEBUG oslo_concurrency.lockutils [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] Acquiring lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.235515] env[63197]: DEBUG oslo_concurrency.lockutils [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.235682] env[63197]: DEBUG oslo_concurrency.lockutils [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.235859] env[63197]: DEBUG nova.compute.manager [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] No waiting events found dispatching network-vif-plugged-720e1317-71ba-4079-a3fb-794d854e0e13 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.237271] env[63197]: WARNING nova.compute.manager [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received unexpected event network-vif-plugged-720e1317-71ba-4079-a3fb-794d854e0e13 for instance with vm_state building and task_state spawning. [ 949.237271] env[63197]: DEBUG nova.compute.manager [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received event network-changed-720e1317-71ba-4079-a3fb-794d854e0e13 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.237271] env[63197]: DEBUG nova.compute.manager [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Refreshing instance network info cache due to event network-changed-720e1317-71ba-4079-a3fb-794d854e0e13. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.237271] env[63197]: DEBUG oslo_concurrency.lockutils [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] Acquiring lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.238201] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955699b2-c775-4068-9cb9-1bda8dd7563a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.258326] env[63197]: DEBUG oslo_vmware.api [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 949.258326] env[63197]: value = "task-1364370" [ 949.258326] env[63197]: _type = "Task" [ 949.258326] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.266420] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cc383e60-474c-4569-bda0-5df97ac2cb7f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.277881] env[63197]: DEBUG oslo_vmware.api [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364370, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.294719] env[63197]: DEBUG oslo_vmware.api [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364369, 'name': PowerOnVM_Task, 'duration_secs': 0.482007} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.294953] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.295262] env[63197]: INFO nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Took 7.18 seconds to spawn the instance on the hypervisor. [ 949.295330] env[63197]: DEBUG nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.296164] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd54e5c-ef14-43c2-8dd4-cf710537d45d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.429162] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0f01ea38-241d-4b6f-aa79-705f666f14fb tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "9c1caa2b-c369-425a-8726-cddadf06f338" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.227s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.430752] env[63197]: DEBUG oslo_concurrency.lockutils [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] Acquired lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.431804] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed5c478-4084-4ceb-a356-88efa6dc603d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.440893] env[63197]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 949.440977] env[63197]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=63197) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 949.441818] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5245ada-a1fe-4c9e-a134-2f28972c84b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.459034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa53db7c-e9ae-4ae0-b58f-606b225292c6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.493224] env[63197]: ERROR root [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-290318' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-290318' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-290318' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-290318'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-290318' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-290318' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-290318'}\n"]: nova.exception.InstanceNotFound: Instance 9c1caa2b-c369-425a-8726-cddadf06f338 could not be found. [ 949.493476] env[63197]: DEBUG oslo_concurrency.lockutils [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] Releasing lock "9c1caa2b-c369-425a-8726-cddadf06f338" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.493651] env[63197]: DEBUG nova.compute.manager [req-8fee3e90-c9db-4003-864e-8a3453aa4ac0 req-23ea940b-241f-4b93-a09c-d9e93abb14a9 service nova] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Detach interface failed, port_id=33a215bc-a3d4-48cf-8856-1da55204dbdf, reason: Instance 9c1caa2b-c369-425a-8726-cddadf06f338 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 949.646502] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affd08fb-2f8d-4ab8-b3a7-b1c71382cffe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.654089] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2962e9be-12a6-4b5d-97e0-88aa598861e0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.687112] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6211a6eb-47b1-44bd-804a-fbb17f0b00d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.694951] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bdd248-36bd-4fbb-845a-f4d227c93e21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.710573] env[63197]: DEBUG nova.compute.provider_tree [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.716397] env[63197]: INFO nova.compute.manager [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] instance snapshotting [ 949.719093] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3aba04b-ba9e-4d92-b653-4a76fc66cd82 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.739548] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379f80a6-4feb-45d1-b3a7-f13d54b55f6f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.768767] env[63197]: DEBUG oslo_vmware.api [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364370, 'name': ReconfigVM_Task, 'duration_secs': 0.204758} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.769164] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290367', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'name': 'volume-07d531bd-d1e0-40e3-8f73-caa2955b6150', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'serial': '07d531bd-d1e0-40e3-8f73-caa2955b6150'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 949.803180] env[63197]: DEBUG nova.network.neutron [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Updating instance_info_cache with network_info: [{"id": "c31363ef-bdd6-49f5-89f7-fb36e583c894", "address": "fa:16:3e:95:3d:14", "network": {"id": "a2d781dd-edcc-4d82-8887-bf588471001d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2103330490", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.170", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc31363ef-bd", "ovs_interfaceid": "c31363ef-bdd6-49f5-89f7-fb36e583c894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "720e1317-71ba-4079-a3fb-794d854e0e13", "address": "fa:16:3e:e9:78:15", "network": {"id": "8eb2a428-eda4-4a68-bb28-a53b00984c1c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-78320906", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap720e1317-71", "ovs_interfaceid": "720e1317-71ba-4079-a3fb-794d854e0e13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.826536] env[63197]: INFO nova.compute.manager [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Took 14.48 seconds to build instance. [ 949.943021] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.943319] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.943521] env[63197]: DEBUG nova.compute.manager [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Going to confirm migration 1 {{(pid=63197) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 950.084495] env[63197]: DEBUG nova.network.neutron [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Successfully updated port: e33a7c35-f308-47b5-8b23-08328d08d9e5 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.215439] env[63197]: DEBUG nova.scheduler.client.report [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.249818] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 950.250830] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9f535798-d356-4c18-949a-563bdf6cd479 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.262402] env[63197]: DEBUG nova.compute.manager [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Received event network-changed-0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.262999] env[63197]: DEBUG nova.compute.manager [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Refreshing instance network info cache due to event network-changed-0bdd3cb1-b50c-4791-9d45-3a1766a307da. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 950.262999] env[63197]: DEBUG oslo_concurrency.lockutils [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] Acquiring lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.263212] env[63197]: DEBUG oslo_concurrency.lockutils [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] Acquired lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.263436] env[63197]: DEBUG nova.network.neutron [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Refreshing network info cache for port 0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.270333] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 950.270333] env[63197]: value = "task-1364371" [ 950.270333] env[63197]: _type = "Task" [ 950.270333] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.284209] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364371, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.306861] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Releasing lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.307019] env[63197]: DEBUG nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Instance network_info: |[{"id": "c31363ef-bdd6-49f5-89f7-fb36e583c894", "address": "fa:16:3e:95:3d:14", "network": {"id": "a2d781dd-edcc-4d82-8887-bf588471001d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2103330490", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.170", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc31363ef-bd", "ovs_interfaceid": "c31363ef-bdd6-49f5-89f7-fb36e583c894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "720e1317-71ba-4079-a3fb-794d854e0e13", "address": "fa:16:3e:e9:78:15", "network": {"id": "8eb2a428-eda4-4a68-bb28-a53b00984c1c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-78320906", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap720e1317-71", "ovs_interfaceid": "720e1317-71ba-4079-a3fb-794d854e0e13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 950.307372] env[63197]: DEBUG oslo_concurrency.lockutils [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] Acquired lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.307565] env[63197]: DEBUG nova.network.neutron [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Refreshing network info cache for port 720e1317-71ba-4079-a3fb-794d854e0e13 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.310290] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:3d:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d054505-89d3-49c5-8b38-5da917a42c49', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c31363ef-bdd6-49f5-89f7-fb36e583c894', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:78:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50886eea-591a-452c-a27b-5f22cfc9df85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '720e1317-71ba-4079-a3fb-794d854e0e13', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.319502] env[63197]: DEBUG oslo.service.loopingcall [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.322797] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 950.323197] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-84a5ad9f-c6b3-4ebd-a90c-117dc77d1398 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.345286] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5b29150c-a4b3-4876-9a4d-739c8e0148f3 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "c6686a09-3711-434b-b69f-a2518366fbed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.999s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.351832] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.351832] env[63197]: value = "task-1364372" [ 950.351832] env[63197]: _type = "Task" [ 950.351832] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.359572] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364372, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.535038] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.536334] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.536701] env[63197]: DEBUG nova.network.neutron [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.537027] env[63197]: DEBUG nova.objects.instance [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lazy-loading 'info_cache' on Instance uuid 537cc8cc-94e0-41b4-8204-6f995f2ea4d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.593359] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-c41b34cb-cc39-461e-83df-cd4de6780d95" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.593359] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-c41b34cb-cc39-461e-83df-cd4de6780d95" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.593359] env[63197]: DEBUG nova.network.neutron [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.723987] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.825s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.728120] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.875s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.731187] env[63197]: INFO nova.compute.claims [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.761617] env[63197]: INFO nova.scheduler.client.report [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Deleted allocations for instance 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6 [ 950.784185] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364371, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.825589] env[63197]: DEBUG nova.objects.instance [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lazy-loading 'flavor' on Instance uuid 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.862526] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364372, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.160282] env[63197]: DEBUG nova.network.neutron [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.275026] env[63197]: DEBUG oslo_concurrency.lockutils [None req-93cd67cb-11ff-461d-b3d1-cde219d765db tempest-ServerRescueTestJSON-198989187 tempest-ServerRescueTestJSON-198989187-project-member] Lock "9c80bc51-c8f6-436d-9d19-47f6d1aadbf6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.857s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.297220] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364371, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.333156] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f512ccf5-0668-451c-94b8-01234dd1c8dd tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.011s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.367260] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364372, 'name': CreateVM_Task, 'duration_secs': 0.519104} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.369349] env[63197]: DEBUG nova.network.neutron [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Updated VIF entry in instance network info cache for port 720e1317-71ba-4079-a3fb-794d854e0e13. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 951.369822] env[63197]: DEBUG nova.network.neutron [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Updating instance_info_cache with network_info: [{"id": "c31363ef-bdd6-49f5-89f7-fb36e583c894", "address": "fa:16:3e:95:3d:14", "network": {"id": "a2d781dd-edcc-4d82-8887-bf588471001d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2103330490", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.170", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d054505-89d3-49c5-8b38-5da917a42c49", "external-id": "nsx-vlan-transportzone-888", "segmentation_id": 888, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc31363ef-bd", "ovs_interfaceid": "c31363ef-bdd6-49f5-89f7-fb36e583c894", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "720e1317-71ba-4079-a3fb-794d854e0e13", "address": "fa:16:3e:e9:78:15", "network": {"id": "8eb2a428-eda4-4a68-bb28-a53b00984c1c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-78320906", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap720e1317-71", "ovs_interfaceid": "720e1317-71ba-4079-a3fb-794d854e0e13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.371483] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 951.376742] env[63197]: DEBUG nova.network.neutron [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updated VIF entry in instance network info cache for port 0bdd3cb1-b50c-4791-9d45-3a1766a307da. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 951.376742] env[63197]: DEBUG nova.network.neutron [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updating instance_info_cache with network_info: [{"id": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "address": "fa:16:3e:49:a5:9e", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bdd3cb1-b5", "ovs_interfaceid": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.380075] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.380075] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.380075] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.383702] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f167117-5a0f-474f-b499-dd739dc3877a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.394155] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 951.394155] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52956ed9-4751-861d-b90c-9bdc3c9284ee" [ 951.394155] env[63197]: _type = "Task" [ 951.394155] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.400462] env[63197]: DEBUG nova.compute.manager [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Received event network-vif-plugged-e33a7c35-f308-47b5-8b23-08328d08d9e5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.400662] env[63197]: DEBUG oslo_concurrency.lockutils [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] Acquiring lock "c41b34cb-cc39-461e-83df-cd4de6780d95-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.400878] env[63197]: DEBUG oslo_concurrency.lockutils [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.401565] env[63197]: DEBUG oslo_concurrency.lockutils [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.401565] env[63197]: DEBUG nova.compute.manager [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] No waiting events found dispatching network-vif-plugged-e33a7c35-f308-47b5-8b23-08328d08d9e5 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.401565] env[63197]: WARNING nova.compute.manager [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Received unexpected event network-vif-plugged-e33a7c35-f308-47b5-8b23-08328d08d9e5 for instance with vm_state building and task_state spawning. [ 951.401565] env[63197]: DEBUG nova.compute.manager [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Received event network-changed-e33a7c35-f308-47b5-8b23-08328d08d9e5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.401825] env[63197]: DEBUG nova.compute.manager [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Refreshing instance network info cache due to event network-changed-e33a7c35-f308-47b5-8b23-08328d08d9e5. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.401883] env[63197]: DEBUG oslo_concurrency.lockutils [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] Acquiring lock "refresh_cache-c41b34cb-cc39-461e-83df-cd4de6780d95" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.409072] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52956ed9-4751-861d-b90c-9bdc3c9284ee, 'name': SearchDatastore_Task, 'duration_secs': 0.013777} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.409498] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.409651] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.409889] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.410977] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.410977] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.410977] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-147775c6-d075-445e-b78d-1ce81a8e2c1e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.419200] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.419493] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 951.423207] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c8a61c8-be20-430a-9085-8e3809dac26c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.425946] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 951.425946] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d01661-ea53-c641-ade6-c8ede970ffb3" [ 951.425946] env[63197]: _type = "Task" [ 951.425946] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.438106] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d01661-ea53-c641-ade6-c8ede970ffb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.454025] env[63197]: DEBUG nova.network.neutron [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Updating instance_info_cache with network_info: [{"id": "e33a7c35-f308-47b5-8b23-08328d08d9e5", "address": "fa:16:3e:dc:a7:6b", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape33a7c35-f3", "ovs_interfaceid": "e33a7c35-f308-47b5-8b23-08328d08d9e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.790302] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364371, 'name': CreateSnapshot_Task, 'duration_secs': 1.105729} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.793538] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 951.794770] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2f129a-3f6a-4e8e-ab10-aaf9c13a5544 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.872683] env[63197]: DEBUG oslo_concurrency.lockutils [req-30153389-19e6-4d51-a471-5cea4fd15408 req-6b4a298a-af4c-4a89-9d4a-5d63b3e1b176 service nova] Releasing lock "refresh_cache-c97f59db-b1bf-4744-98ed-b6d7c095207f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.875397] env[63197]: DEBUG oslo_concurrency.lockutils [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.875667] env[63197]: DEBUG oslo_concurrency.lockutils [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.884500] env[63197]: DEBUG oslo_concurrency.lockutils [req-736a057e-911c-4fc6-9367-e947f09e566f req-4f18bcb6-9752-4b21-bae0-6eac1b2a3018 service nova] Releasing lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.944286] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d01661-ea53-c641-ade6-c8ede970ffb3, 'name': SearchDatastore_Task, 'duration_secs': 0.012406} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.945293] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cde05677-574e-4da5-a3ac-204bf3819d9c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.952360] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2f40d70-826a-4bf2-9c19-63382f3a7a61 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.957582] env[63197]: DEBUG nova.network.neutron [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance_info_cache with network_info: [{"id": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "address": "fa:16:3e:be:d6:04", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d53dd8b-1b", "ovs_interfaceid": "2d53dd8b-1ba1-495d-99fc-265d17cb34a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.963146] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-c41b34cb-cc39-461e-83df-cd4de6780d95" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.963146] env[63197]: DEBUG nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Instance network_info: |[{"id": "e33a7c35-f308-47b5-8b23-08328d08d9e5", "address": "fa:16:3e:dc:a7:6b", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape33a7c35-f3", "ovs_interfaceid": "e33a7c35-f308-47b5-8b23-08328d08d9e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.963146] env[63197]: DEBUG oslo_concurrency.lockutils [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] Acquired lock "refresh_cache-c41b34cb-cc39-461e-83df-cd4de6780d95" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.963146] env[63197]: DEBUG nova.network.neutron [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Refreshing network info cache for port e33a7c35-f308-47b5-8b23-08328d08d9e5 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.964125] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:a7:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e33a7c35-f308-47b5-8b23-08328d08d9e5', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.972065] env[63197]: DEBUG oslo.service.loopingcall [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.974470] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.975046] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 951.975046] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d15a0e-8d96-596b-519f-5411e0d4ad55" [ 951.975046] env[63197]: _type = "Task" [ 951.975046] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.977293] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2008dfca-2791-4d2c-a7cb-5b8c39d88360 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.003438] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e031c7b9-9c17-4001-a723-edc930549566 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.017750] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d15a0e-8d96-596b-519f-5411e0d4ad55, 'name': SearchDatastore_Task, 'duration_secs': 0.011385} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.043897] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.044222] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] c97f59db-b1bf-4744-98ed-b6d7c095207f/c97f59db-b1bf-4744-98ed-b6d7c095207f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 952.045066] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.045066] env[63197]: value = "task-1364373" [ 952.045066] env[63197]: _type = "Task" [ 952.045066] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.045303] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b1aa85c-00e7-4349-beaa-e5223ed9fbbf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.047747] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8f50c3-4d21-4dcb-bd8f-4415aa907036 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.061779] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00a5f9e-914a-4355-aefc-acd6654fb6a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.065668] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364373, 'name': CreateVM_Task} progress is 15%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.066022] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 952.066022] env[63197]: value = "task-1364374" [ 952.066022] env[63197]: _type = "Task" [ 952.066022] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.077994] env[63197]: DEBUG nova.compute.provider_tree [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.084381] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364374, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.319541] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 952.321574] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-67e04855-2303-4b20-931b-08adc072633d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.330906] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 952.330906] env[63197]: value = "task-1364375" [ 952.330906] env[63197]: _type = "Task" [ 952.330906] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.340839] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364375, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.379846] env[63197]: INFO nova.compute.manager [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Detaching volume c29df051-5e11-4357-9d18-1c7dc22c901b [ 952.438183] env[63197]: INFO nova.virt.block_device [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Attempting to driver detach volume c29df051-5e11-4357-9d18-1c7dc22c901b from mountpoint /dev/sdb [ 952.438605] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 952.438843] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290360', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'name': 'volume-c29df051-5e11-4357-9d18-1c7dc22c901b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'serial': 'c29df051-5e11-4357-9d18-1c7dc22c901b'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 952.440675] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd393b9-7cb5-4d44-a112-b151e8a09780 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.470741] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-537cc8cc-94e0-41b4-8204-6f995f2ea4d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.471089] env[63197]: DEBUG nova.objects.instance [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lazy-loading 'migration_context' on Instance uuid 537cc8cc-94e0-41b4-8204-6f995f2ea4d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.473115] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6e36ad-fcb0-4ff6-9f31-2df3dbc50d4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.489459] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853c6f9d-d465-4300-9c0b-aa484562bba5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.523246] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceafd192-bc7c-46d0-b3da-d89191c67c12 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.549179] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] The volume has not been displaced from its original location: [datastore2] volume-c29df051-5e11-4357-9d18-1c7dc22c901b/volume-c29df051-5e11-4357-9d18-1c7dc22c901b.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 952.554579] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfiguring VM instance instance-00000044 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 952.555101] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fd3b5a7-ec4d-4f6c-9837-b14837640dd5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.584671] env[63197]: DEBUG nova.scheduler.client.report [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.588823] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364373, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.595942] env[63197]: DEBUG oslo_vmware.api [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 952.595942] env[63197]: value = "task-1364376" [ 952.595942] env[63197]: _type = "Task" [ 952.595942] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.595942] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364374, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.609299] env[63197]: DEBUG oslo_vmware.api [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364376, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.653896] env[63197]: DEBUG nova.compute.manager [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Received event network-changed-0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.653896] env[63197]: DEBUG nova.compute.manager [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Refreshing instance network info cache due to event network-changed-0bdd3cb1-b50c-4791-9d45-3a1766a307da. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.653896] env[63197]: DEBUG oslo_concurrency.lockutils [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] Acquiring lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.653896] env[63197]: DEBUG oslo_concurrency.lockutils [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] Acquired lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.654221] env[63197]: DEBUG nova.network.neutron [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Refreshing network info cache for port 0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.841745] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364375, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.891058] env[63197]: DEBUG nova.network.neutron [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Updated VIF entry in instance network info cache for port e33a7c35-f308-47b5-8b23-08328d08d9e5. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.891225] env[63197]: DEBUG nova.network.neutron [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Updating instance_info_cache with network_info: [{"id": "e33a7c35-f308-47b5-8b23-08328d08d9e5", "address": "fa:16:3e:dc:a7:6b", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape33a7c35-f3", "ovs_interfaceid": "e33a7c35-f308-47b5-8b23-08328d08d9e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.979320] env[63197]: DEBUG nova.objects.base [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Object Instance<537cc8cc-94e0-41b4-8204-6f995f2ea4d3> lazy-loaded attributes: info_cache,migration_context {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 952.980403] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2384546-693e-4a1d-9e09-604c49fba85d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.005206] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54c5452a-c871-4946-815d-550d8e942d73 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.011710] env[63197]: DEBUG oslo_vmware.api [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 953.011710] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5275e771-0fbc-c115-e716-14200390d8fc" [ 953.011710] env[63197]: _type = "Task" [ 953.011710] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.020027] env[63197]: DEBUG oslo_vmware.api [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5275e771-0fbc-c115-e716-14200390d8fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.065400] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364373, 'name': CreateVM_Task, 'duration_secs': 0.633461} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.065786] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.066401] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.066572] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.066901] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.067185] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c7a31e2-fefa-4ebd-a0af-cb4a8371ab94 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.072071] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 953.072071] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5221bb7a-a323-7875-2bbc-94c5a8a861fd" [ 953.072071] env[63197]: _type = "Task" [ 953.072071] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.091864] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364374, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585445} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.095820] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.096574] env[63197]: DEBUG nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.099414] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] c97f59db-b1bf-4744-98ed-b6d7c095207f/c97f59db-b1bf-4744-98ed-b6d7c095207f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.099648] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.099923] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5221bb7a-a323-7875-2bbc-94c5a8a861fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.100443] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f1c84243-f49b-4b0a-87a9-9dcb9b9c65e9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.110883] env[63197]: DEBUG oslo_vmware.api [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364376, 'name': ReconfigVM_Task, 'duration_secs': 0.34682} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.112320] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfigured VM instance instance-00000044 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.117523] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 953.117523] env[63197]: value = "task-1364377" [ 953.117523] env[63197]: _type = "Task" [ 953.117523] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.117810] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19c8aa56-dd34-430c-b6ee-d32e3f5ca8bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.137894] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.139418] env[63197]: DEBUG oslo_vmware.api [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 953.139418] env[63197]: value = "task-1364378" [ 953.139418] env[63197]: _type = "Task" [ 953.139418] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.148741] env[63197]: DEBUG oslo_vmware.api [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364378, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.203022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "4672f595-e1f6-4400-b5a1-065598584980" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.203022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "4672f595-e1f6-4400-b5a1-065598584980" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.345295] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364375, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.394900] env[63197]: DEBUG oslo_concurrency.lockutils [req-b5928324-67e2-41a0-b55c-dddef7098041 req-b148ce4d-78ae-4d52-861d-f333afe79d7c service nova] Releasing lock "refresh_cache-c41b34cb-cc39-461e-83df-cd4de6780d95" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.479607] env[63197]: DEBUG nova.network.neutron [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updated VIF entry in instance network info cache for port 0bdd3cb1-b50c-4791-9d45-3a1766a307da. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.479607] env[63197]: DEBUG nova.network.neutron [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updating instance_info_cache with network_info: [{"id": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "address": "fa:16:3e:49:a5:9e", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0bdd3cb1-b5", "ovs_interfaceid": "0bdd3cb1-b50c-4791-9d45-3a1766a307da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.525514] env[63197]: DEBUG oslo_vmware.api [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5275e771-0fbc-c115-e716-14200390d8fc, 'name': SearchDatastore_Task, 'duration_secs': 0.012746} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.525767] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.526016] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.585644] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5221bb7a-a323-7875-2bbc-94c5a8a861fd, 'name': SearchDatastore_Task, 'duration_secs': 0.015764} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.585942] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.586197] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.586438] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.586583] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.586767] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.587048] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53e08e83-4444-40cc-ba14-2e777f39c4e3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.599752] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.599992] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.601738] env[63197]: DEBUG nova.compute.utils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.603060] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f0b4327-dfa6-4878-820f-440ccc2d83fd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.606029] env[63197]: DEBUG nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.606130] env[63197]: DEBUG nova.network.neutron [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 953.610933] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 953.610933] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52859ca3-8eca-0ea0-ae93-d1981035472d" [ 953.610933] env[63197]: _type = "Task" [ 953.610933] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.619078] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52859ca3-8eca-0ea0-ae93-d1981035472d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.638547] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364377, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084117} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.638823] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 953.639603] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccca450-b7fb-4282-876b-953f7d160736 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.667976] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] c97f59db-b1bf-4744-98ed-b6d7c095207f/c97f59db-b1bf-4744-98ed-b6d7c095207f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 953.670021] env[63197]: DEBUG nova.policy [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '514b7795ecdc49069e7f3cab700a0ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '527721557235413e99c4a2eaa086486c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.674583] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-101243b1-dabc-4180-a4ee-1359b0bd6f21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.688745] env[63197]: DEBUG oslo_vmware.api [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364378, 'name': ReconfigVM_Task, 'duration_secs': 0.168928} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.689594] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290360', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'name': 'volume-c29df051-5e11-4357-9d18-1c7dc22c901b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': 'c29df051-5e11-4357-9d18-1c7dc22c901b', 'serial': 'c29df051-5e11-4357-9d18-1c7dc22c901b'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 953.697521] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 953.697521] env[63197]: value = "task-1364379" [ 953.697521] env[63197]: _type = "Task" [ 953.697521] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.706851] env[63197]: DEBUG nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.709760] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364379, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.843844] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364375, 'name': CloneVM_Task} progress is 95%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.981310] env[63197]: DEBUG oslo_concurrency.lockutils [req-1e528195-a85d-4797-ae2c-d98d125e5f85 req-14258922-399f-492d-a5d2-1e302b1659ed service nova] Releasing lock "refresh_cache-c6686a09-3711-434b-b69f-a2518366fbed" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.023779] env[63197]: DEBUG nova.network.neutron [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Successfully created port: fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.107849] env[63197]: DEBUG nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.125796] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52859ca3-8eca-0ea0-ae93-d1981035472d, 'name': SearchDatastore_Task, 'duration_secs': 0.022174} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.130898] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b33e46c-2cd5-43a9-b349-1dcb7058a0e2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.137293] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 954.137293] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525d4a31-44fe-f1a0-b47c-3ab597e2a6a3" [ 954.137293] env[63197]: _type = "Task" [ 954.137293] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.148226] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525d4a31-44fe-f1a0-b47c-3ab597e2a6a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.205346] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8f6936-b0cf-42de-9026-06e08dd2b0c9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.215684] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364379, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.221912] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d277f83-6770-48f2-a4d6-22e675d6e27b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.227781] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.254887] env[63197]: DEBUG nova.objects.instance [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lazy-loading 'flavor' on Instance uuid 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.257920] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48573b29-ac98-4d1b-ade2-3d791a5c7dc1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.264388] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61e8226-b1b5-4843-af9f-4ead32a1adf0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.279058] env[63197]: DEBUG nova.compute.provider_tree [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.345769] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364375, 'name': CloneVM_Task, 'duration_secs': 1.67172} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.349021] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Created linked-clone VM from snapshot [ 954.349021] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac311bb1-fe0d-4a56-8a57-92c47d9c91b6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.358313] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Uploading image 0b7db7fb-059d-491c-979a-4ff6763069f5 {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 954.388621] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 954.388621] env[63197]: value = "vm-290376" [ 954.388621] env[63197]: _type = "VirtualMachine" [ 954.388621] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 954.388985] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-06b82ca6-8e77-44b3-aaef-09abc155d1e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.396636] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lease: (returnval){ [ 954.396636] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a958bc-2c03-e0f9-cd19-d7ce5912bffe" [ 954.396636] env[63197]: _type = "HttpNfcLease" [ 954.396636] env[63197]: } obtained for exporting VM: (result){ [ 954.396636] env[63197]: value = "vm-290376" [ 954.396636] env[63197]: _type = "VirtualMachine" [ 954.396636] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 954.396987] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the lease: (returnval){ [ 954.396987] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a958bc-2c03-e0f9-cd19-d7ce5912bffe" [ 954.396987] env[63197]: _type = "HttpNfcLease" [ 954.396987] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 954.404073] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.404073] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a958bc-2c03-e0f9-cd19-d7ce5912bffe" [ 954.404073] env[63197]: _type = "HttpNfcLease" [ 954.404073] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 954.647714] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525d4a31-44fe-f1a0-b47c-3ab597e2a6a3, 'name': SearchDatastore_Task, 'duration_secs': 0.024299} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.648018] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.648304] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] c41b34cb-cc39-461e-83df-cd4de6780d95/c41b34cb-cc39-461e-83df-cd4de6780d95.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.648578] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f98e4c9-6e7d-4622-bdd6-42035a29028a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.656046] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 954.656046] env[63197]: value = "task-1364381" [ 954.656046] env[63197]: _type = "Task" [ 954.656046] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.666221] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.692781] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.692781] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.707455] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364379, 'name': ReconfigVM_Task, 'duration_secs': 0.729324} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.707739] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Reconfigured VM instance instance-00000052 to attach disk [datastore1] c97f59db-b1bf-4744-98ed-b6d7c095207f/c97f59db-b1bf-4744-98ed-b6d7c095207f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.708503] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2324b51d-7008-40f0-840e-ed050e07fdb7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.716335] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 954.716335] env[63197]: value = "task-1364382" [ 954.716335] env[63197]: _type = "Task" [ 954.716335] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.725515] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364382, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.782483] env[63197]: DEBUG nova.scheduler.client.report [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.906310] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.906310] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a958bc-2c03-e0f9-cd19-d7ce5912bffe" [ 954.906310] env[63197]: _type = "HttpNfcLease" [ 954.906310] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.906763] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.906763] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a958bc-2c03-e0f9-cd19-d7ce5912bffe" [ 954.906763] env[63197]: _type = "HttpNfcLease" [ 954.906763] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 954.907856] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44bcfc5b-8450-44fb-85f9-4c4d52aac82a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.915253] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c806bb-7df4-23d9-5d07-969dc9945cc5/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.915488] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c806bb-7df4-23d9-5d07-969dc9945cc5/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 955.076508] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-378247e9-6549-4c05-b7e1-f8112fd9e742 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.122814] env[63197]: DEBUG nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.166810] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364381, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.198086] env[63197]: DEBUG nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 955.233143] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.233731] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364382, 'name': Rename_Task, 'duration_secs': 0.216145} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.234165] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 955.235371] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45636c2e-c828-463a-ba00-63b31ebafb8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.246873] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 955.246873] env[63197]: value = "task-1364383" [ 955.246873] env[63197]: _type = "Task" [ 955.246873] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.260133] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364383, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.265207] env[63197]: DEBUG oslo_concurrency.lockutils [None req-00b96b7a-0d93-4486-910f-ef4ef31d4ab5 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.389s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.266931] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.034s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.328922] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.329229] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.329509] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.329706] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.329858] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.330064] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.330469] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.330761] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.331068] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.331250] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.331432] env[63197]: DEBUG nova.virt.hardware [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.332684] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89e288f-72b8-4b4f-8a35-2b7d8733505b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.341056] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c5e3cc3-58ce-4f32-96b6-5d4bde3d3378 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.605547] env[63197]: DEBUG nova.compute.manager [req-71091207-db21-4bf7-8b39-c649b5589544 req-f4036f7a-180b-4223-ba3e-5730df474cc9 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Received event network-vif-plugged-fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.606186] env[63197]: DEBUG oslo_concurrency.lockutils [req-71091207-db21-4bf7-8b39-c649b5589544 req-f4036f7a-180b-4223-ba3e-5730df474cc9 service nova] Acquiring lock "8e5ef1b0-7532-498a-84c2-189274a36c50-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.607311] env[63197]: DEBUG oslo_concurrency.lockutils [req-71091207-db21-4bf7-8b39-c649b5589544 req-f4036f7a-180b-4223-ba3e-5730df474cc9 service nova] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.607583] env[63197]: DEBUG oslo_concurrency.lockutils [req-71091207-db21-4bf7-8b39-c649b5589544 req-f4036f7a-180b-4223-ba3e-5730df474cc9 service nova] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.607854] env[63197]: DEBUG nova.compute.manager [req-71091207-db21-4bf7-8b39-c649b5589544 req-f4036f7a-180b-4223-ba3e-5730df474cc9 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] No waiting events found dispatching network-vif-plugged-fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.608126] env[63197]: WARNING nova.compute.manager [req-71091207-db21-4bf7-8b39-c649b5589544 req-f4036f7a-180b-4223-ba3e-5730df474cc9 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Received unexpected event network-vif-plugged-fa39c111-5e72-43e8-b764-d2072c0c3d40 for instance with vm_state building and task_state spawning. [ 955.668054] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655435} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.668453] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] c41b34cb-cc39-461e-83df-cd4de6780d95/c41b34cb-cc39-461e-83df-cd4de6780d95.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.668453] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.668799] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86286356-5f6e-474d-a1b4-8b2027bdb278 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.677446] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 955.677446] env[63197]: value = "task-1364384" [ 955.677446] env[63197]: _type = "Task" [ 955.677446] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.685947] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364384, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.728926] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.736870] env[63197]: DEBUG nova.network.neutron [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Successfully updated port: fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.759436] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364383, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.774264] env[63197]: INFO nova.compute.manager [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Detaching volume 07d531bd-d1e0-40e3-8f73-caa2955b6150 [ 955.797572] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.271s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.802922] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.576s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.806491] env[63197]: INFO nova.compute.claims [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.811576] env[63197]: INFO nova.virt.block_device [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Attempting to driver detach volume 07d531bd-d1e0-40e3-8f73-caa2955b6150 from mountpoint /dev/sdc [ 955.811880] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 955.812188] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290367', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'name': 'volume-07d531bd-d1e0-40e3-8f73-caa2955b6150', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'serial': '07d531bd-d1e0-40e3-8f73-caa2955b6150'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 955.813889] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075a825b-5689-4404-b6b0-5cfa0df5c238 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.839117] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1cbcc4-9490-4712-941c-bd2977f2fec2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.847782] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654ed7b8-7e4d-45c5-b880-b468c3877892 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.872094] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d61908-3067-4e70-8f12-3d2f3e3b2355 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.891017] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] The volume has not been displaced from its original location: [datastore1] volume-07d531bd-d1e0-40e3-8f73-caa2955b6150/volume-07d531bd-d1e0-40e3-8f73-caa2955b6150.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 955.896567] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfiguring VM instance instance-00000044 to detach disk 2002 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 955.896962] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cf3e283-4515-4b3e-be69-971fdd5523b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.915933] env[63197]: DEBUG oslo_vmware.api [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 955.915933] env[63197]: value = "task-1364385" [ 955.915933] env[63197]: _type = "Task" [ 955.915933] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.926156] env[63197]: DEBUG oslo_vmware.api [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364385, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.187535] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364384, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069515} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.187978] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.188879] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf315fc-0f9a-4789-9190-9023b3591862 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.213944] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] c41b34cb-cc39-461e-83df-cd4de6780d95/c41b34cb-cc39-461e-83df-cd4de6780d95.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.214444] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e9ac76d-514f-4455-92b7-0dd26d9de063 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.234240] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 956.234240] env[63197]: value = "task-1364386" [ 956.234240] env[63197]: _type = "Task" [ 956.234240] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.243284] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.243543] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.243731] env[63197]: DEBUG nova.network.neutron [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.244978] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364386, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.258113] env[63197]: DEBUG oslo_vmware.api [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364383, 'name': PowerOnVM_Task, 'duration_secs': 0.700113} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.258629] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.258938] env[63197]: INFO nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Took 11.41 seconds to spawn the instance on the hypervisor. [ 956.259230] env[63197]: DEBUG nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 956.260240] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b001dd7e-e754-4d70-832f-4bacfe9cc902 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.379177] env[63197]: INFO nova.scheduler.client.report [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted allocation for migration 90c18999-4dd5-4787-8836-75a6a0374213 [ 956.431522] env[63197]: DEBUG oslo_vmware.api [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364385, 'name': ReconfigVM_Task, 'duration_secs': 0.312051} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.433556] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Reconfigured VM instance instance-00000044 to detach disk 2002 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 956.442742] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e3e556a-5c8a-4d6d-8a60-0512d047779e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.470724] env[63197]: DEBUG oslo_vmware.api [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 956.470724] env[63197]: value = "task-1364387" [ 956.470724] env[63197]: _type = "Task" [ 956.470724] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.484292] env[63197]: DEBUG oslo_vmware.api [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364387, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.748054] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364386, 'name': ReconfigVM_Task, 'duration_secs': 0.458806} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.748054] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Reconfigured VM instance instance-00000053 to attach disk [datastore1] c41b34cb-cc39-461e-83df-cd4de6780d95/c41b34cb-cc39-461e-83df-cd4de6780d95.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.750663] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-20ff89da-15d5-4534-9e3a-216df1685fd6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.758240] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 956.758240] env[63197]: value = "task-1364388" [ 956.758240] env[63197]: _type = "Task" [ 956.758240] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.770714] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364388, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.786228] env[63197]: INFO nova.compute.manager [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Took 20.46 seconds to build instance. [ 956.796423] env[63197]: DEBUG nova.network.neutron [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.889022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-552b8859-3b5a-4266-b3fb-30626d3c8090 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.945s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.984433] env[63197]: DEBUG oslo_vmware.api [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364387, 'name': ReconfigVM_Task, 'duration_secs': 0.155343} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.987865] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290367', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'name': 'volume-07d531bd-d1e0-40e3-8f73-caa2955b6150', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5b9c6a1c-4212-4941-aa6c-364fb82f5e64', 'attached_at': '', 'detached_at': '', 'volume_id': '07d531bd-d1e0-40e3-8f73-caa2955b6150', 'serial': '07d531bd-d1e0-40e3-8f73-caa2955b6150'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 957.038397] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f85dab0-5139-4de7-b131-abf2eca0a8c6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.048771] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5d1dfb6-e60e-4683-bc23-6e45d19feda0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.054753] env[63197]: DEBUG nova.network.neutron [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Updating instance_info_cache with network_info: [{"id": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "address": "fa:16:3e:86:e6:7f", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa39c111-5e", "ovs_interfaceid": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.089382] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cc9687-6aab-428a-89a1-a571363d71b7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.098854] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9336ade-c783-417a-9ea6-fd4d37044646 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.114197] env[63197]: DEBUG nova.compute.provider_tree [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.270288] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364388, 'name': Rename_Task, 'duration_secs': 0.192818} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.270466] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.270641] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66e4a82f-98d7-4abb-a7e4-2904135ce80b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.276551] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 957.276551] env[63197]: value = "task-1364389" [ 957.276551] env[63197]: _type = "Task" [ 957.276551] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.285515] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364389, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.290510] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2fe529be-8231-4ad4-9041-9e27ef823c60 tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.978s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.518379] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "c97f59db-b1bf-4744-98ed-b6d7c095207f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.518685] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.518905] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.519105] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.519316] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.521611] env[63197]: INFO nova.compute.manager [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Terminating instance [ 957.523777] env[63197]: DEBUG nova.compute.manager [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.523977] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.524880] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecc5582-8317-4183-afeb-1115a31543b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.534040] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.534040] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec6c5563-64b5-41b1-88b7-03c8b811b7de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.540722] env[63197]: DEBUG oslo_vmware.api [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 957.540722] env[63197]: value = "task-1364390" [ 957.540722] env[63197]: _type = "Task" [ 957.540722] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.546022] env[63197]: DEBUG nova.objects.instance [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lazy-loading 'flavor' on Instance uuid 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.551660] env[63197]: DEBUG oslo_vmware.api [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364390, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.557730] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.558071] env[63197]: DEBUG nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Instance network_info: |[{"id": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "address": "fa:16:3e:86:e6:7f", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa39c111-5e", "ovs_interfaceid": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.558539] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:e6:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa39c111-5e72-43e8-b764-d2072c0c3d40', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.566425] env[63197]: DEBUG oslo.service.loopingcall [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.567702] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.567971] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cad17da-dac1-4c56-af89-aa70c0a7e5a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.589465] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.589465] env[63197]: value = "task-1364391" [ 957.589465] env[63197]: _type = "Task" [ 957.589465] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.599395] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364391, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.618784] env[63197]: DEBUG nova.scheduler.client.report [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.787618] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364389, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.994802] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.995156] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.995209] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.995385] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.995567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.997849] env[63197]: INFO nova.compute.manager [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Terminating instance [ 957.999796] env[63197]: DEBUG nova.compute.manager [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 958.000036] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.000893] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78fca4b-6e07-42b9-afd5-6a7072194610 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.009340] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.009636] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7c2f406-03d4-49ee-b1b3-667cbad0a0fe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.016231] env[63197]: DEBUG oslo_vmware.api [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 958.016231] env[63197]: value = "task-1364392" [ 958.016231] env[63197]: _type = "Task" [ 958.016231] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.027295] env[63197]: DEBUG oslo_vmware.api [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364392, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.051240] env[63197]: DEBUG oslo_vmware.api [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364390, 'name': PowerOffVM_Task, 'duration_secs': 0.215851} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.051508] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.051711] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.051981] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58b5f4bc-7024-4381-ae8d-37c80737b9a1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.101462] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364391, 'name': CreateVM_Task, 'duration_secs': 0.377823} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.102069] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.102905] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.103103] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.103491] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.103771] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfddcfc5-3568-4fb1-8112-028e3daffb18 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.108872] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 958.108872] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521ee0b2-f8cc-19dc-6185-1503158b5576" [ 958.108872] env[63197]: _type = "Task" [ 958.108872] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.119811] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521ee0b2-f8cc-19dc-6185-1503158b5576, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.124842] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.125407] env[63197]: DEBUG nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 958.128508] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.400s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.131302] env[63197]: INFO nova.compute.claims [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.178456] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.179153] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.179153] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Deleting the datastore file [datastore1] c97f59db-b1bf-4744-98ed-b6d7c095207f {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.180458] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20628028-a6e9-4cfa-a989-5cd70e728956 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.189174] env[63197]: DEBUG oslo_vmware.api [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for the task: (returnval){ [ 958.189174] env[63197]: value = "task-1364394" [ 958.189174] env[63197]: _type = "Task" [ 958.189174] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.198934] env[63197]: DEBUG oslo_vmware.api [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364394, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.287892] env[63197]: DEBUG oslo_vmware.api [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364389, 'name': PowerOnVM_Task, 'duration_secs': 0.628211} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.288205] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.288416] env[63197]: INFO nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Took 9.18 seconds to spawn the instance on the hypervisor. [ 958.288599] env[63197]: DEBUG nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.289422] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b20810-1592-4d1b-9d2f-0751e0160587 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.526698] env[63197]: DEBUG oslo_vmware.api [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364392, 'name': PowerOffVM_Task, 'duration_secs': 0.197165} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.527044] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.527224] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.527493] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1fe2f9b-734f-4cce-9d39-7e5753d253d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.559755] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7750b801-4b67-4a60-9b58-1bb0e8370bee tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.293s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.602503] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.602759] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.602949] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleting the datastore file [datastore1] 537cc8cc-94e0-41b4-8204-6f995f2ea4d3 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.603246] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b665620-7c95-4ddb-9cbc-81a3c37bb1a6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.610449] env[63197]: DEBUG oslo_vmware.api [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 958.610449] env[63197]: value = "task-1364396" [ 958.610449] env[63197]: _type = "Task" [ 958.610449] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.626605] env[63197]: DEBUG oslo_vmware.api [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364396, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.627026] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521ee0b2-f8cc-19dc-6185-1503158b5576, 'name': SearchDatastore_Task, 'duration_secs': 0.013206} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.627398] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.627782] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.628136] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.628379] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.628667] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.629024] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc58fd09-9d05-4f27-a5bc-d503fde94268 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.636469] env[63197]: DEBUG nova.compute.utils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.642015] env[63197]: DEBUG nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 958.642331] env[63197]: DEBUG nova.network.neutron [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 958.646119] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.646119] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 958.647111] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3977b231-bb6a-4fe0-b1ec-df8065c95861 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.656389] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 958.656389] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bd38d3-1154-1b6d-c68b-ccfad178d69f" [ 958.656389] env[63197]: _type = "Task" [ 958.656389] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.665853] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bd38d3-1154-1b6d-c68b-ccfad178d69f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.691017] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ebcd2-f6e3-f04a-e49d-281a1a6e1d2d/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 958.692404] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384351af-7332-43ab-949b-ada9f87835c4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.703927] env[63197]: DEBUG oslo_vmware.api [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Task: {'id': task-1364394, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278704} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.706728] env[63197]: DEBUG nova.policy [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.708374] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.708600] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.708798] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.709010] env[63197]: INFO nova.compute.manager [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Took 1.19 seconds to destroy the instance on the hypervisor. [ 958.709263] env[63197]: DEBUG oslo.service.loopingcall [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.709541] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ebcd2-f6e3-f04a-e49d-281a1a6e1d2d/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 958.709734] env[63197]: ERROR oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ebcd2-f6e3-f04a-e49d-281a1a6e1d2d/disk-0.vmdk due to incomplete transfer. [ 958.710982] env[63197]: DEBUG nova.compute.manager [-] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 958.711113] env[63197]: DEBUG nova.network.neutron [-] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 958.712708] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6c35c56c-e154-4330-9511-f2606498cf98 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.715676] env[63197]: DEBUG nova.compute.manager [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Received event network-changed-fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.715893] env[63197]: DEBUG nova.compute.manager [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Refreshing instance network info cache due to event network-changed-fa39c111-5e72-43e8-b764-d2072c0c3d40. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.716185] env[63197]: DEBUG oslo_concurrency.lockutils [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] Acquiring lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.717052] env[63197]: DEBUG oslo_concurrency.lockutils [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] Acquired lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.717052] env[63197]: DEBUG nova.network.neutron [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Refreshing network info cache for port fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.724835] env[63197]: DEBUG oslo_vmware.rw_handles [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527ebcd2-f6e3-f04a-e49d-281a1a6e1d2d/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 958.725060] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Uploaded image 073c5ea5-cd55-4c9a-ab31-9918df041ed1 to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 958.727365] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 958.728423] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-0e7ca96c-8635-4c13-94b5-a909c34ab12d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.735633] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 958.735633] env[63197]: value = "task-1364397" [ 958.735633] env[63197]: _type = "Task" [ 958.735633] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.745066] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364397, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.810670] env[63197]: INFO nova.compute.manager [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Took 18.15 seconds to build instance. [ 959.124525] env[63197]: DEBUG oslo_vmware.api [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364396, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.310699} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.124828] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.125273] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.125538] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.125719] env[63197]: INFO nova.compute.manager [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 959.125964] env[63197]: DEBUG oslo.service.loopingcall [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.126190] env[63197]: DEBUG nova.compute.manager [-] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.126285] env[63197]: DEBUG nova.network.neutron [-] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 959.143203] env[63197]: DEBUG nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 959.167917] env[63197]: DEBUG nova.network.neutron [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Successfully created port: 6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 959.176910] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bd38d3-1154-1b6d-c68b-ccfad178d69f, 'name': SearchDatastore_Task, 'duration_secs': 0.015862} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.180314] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e510ffef-9943-40f8-a6cb-f9b795ea66be {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.186261] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 959.186261] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52749305-966b-704d-0752-928e6fab7cff" [ 959.186261] env[63197]: _type = "Task" [ 959.186261] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.200741] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52749305-966b-704d-0752-928e6fab7cff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.248965] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364397, 'name': Destroy_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.315403] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3d1a3aaa-3568-4f1d-9e81-5b6563c572f3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.666s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.340517] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700a0b19-bff9-4edd-8968-c6ab52d9d156 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.352259] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257d55fd-33a2-4eb0-842c-dc5b141af0f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.383319] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37664bb5-22d8-478a-8a6a-cd9826bc516e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.392020] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390248e8-a00d-4511-befc-45e70f6b46a6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.406164] env[63197]: DEBUG nova.compute.provider_tree [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.564423] env[63197]: DEBUG nova.compute.manager [req-202aac7b-7f65-40c0-bec3-6bdd8ec3bf59 req-03f87671-3d1f-40a5-80fc-be4a8532b43c service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received event network-vif-deleted-c31363ef-bdd6-49f5-89f7-fb36e583c894 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.564695] env[63197]: INFO nova.compute.manager [req-202aac7b-7f65-40c0-bec3-6bdd8ec3bf59 req-03f87671-3d1f-40a5-80fc-be4a8532b43c service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Neutron deleted interface c31363ef-bdd6-49f5-89f7-fb36e583c894; detaching it from the instance and deleting it from the info cache [ 959.564927] env[63197]: DEBUG nova.network.neutron [req-202aac7b-7f65-40c0-bec3-6bdd8ec3bf59 req-03f87671-3d1f-40a5-80fc-be4a8532b43c service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Updating instance_info_cache with network_info: [{"id": "720e1317-71ba-4079-a3fb-794d854e0e13", "address": "fa:16:3e:e9:78:15", "network": {"id": "8eb2a428-eda4-4a68-bb28-a53b00984c1c", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-78320906", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e4f082d2f82c42218116efa35aee1f56", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap720e1317-71", "ovs_interfaceid": "720e1317-71ba-4079-a3fb-794d854e0e13", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.705510] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52749305-966b-704d-0752-928e6fab7cff, 'name': SearchDatastore_Task, 'duration_secs': 0.014078} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.710950] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.711267] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 8e5ef1b0-7532-498a-84c2-189274a36c50/8e5ef1b0-7532-498a-84c2-189274a36c50.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.713545] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "64b944ea-2397-45eb-a0ed-9f57263db998" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.713765] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.718270] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9ebdb02-15ea-402e-9322-80a0d6cd836f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.727340] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 959.727340] env[63197]: value = "task-1364398" [ 959.727340] env[63197]: _type = "Task" [ 959.727340] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.736670] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364398, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.749012] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364397, 'name': Destroy_Task, 'duration_secs': 0.645843} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.749012] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Destroyed the VM [ 959.749221] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 959.749496] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0064521b-ca13-49fc-922f-6d587d384ad9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.756204] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 959.756204] env[63197]: value = "task-1364399" [ 959.756204] env[63197]: _type = "Task" [ 959.756204] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.764342] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364399, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.771383] env[63197]: DEBUG nova.network.neutron [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Updated VIF entry in instance network info cache for port fa39c111-5e72-43e8-b764-d2072c0c3d40. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.771817] env[63197]: DEBUG nova.network.neutron [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Updating instance_info_cache with network_info: [{"id": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "address": "fa:16:3e:86:e6:7f", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa39c111-5e", "ovs_interfaceid": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.909513] env[63197]: DEBUG nova.scheduler.client.report [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.069278] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f610ff73-85ac-46b0-aa24-aea49ac0891d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.079864] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d81ccca-4771-4392-9cef-d3815cedcfc1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.114783] env[63197]: DEBUG nova.compute.manager [req-202aac7b-7f65-40c0-bec3-6bdd8ec3bf59 req-03f87671-3d1f-40a5-80fc-be4a8532b43c service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Detach interface failed, port_id=c31363ef-bdd6-49f5-89f7-fb36e583c894, reason: Instance c97f59db-b1bf-4744-98ed-b6d7c095207f could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 960.132599] env[63197]: DEBUG nova.network.neutron [-] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.154985] env[63197]: DEBUG nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 960.181931] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 960.182281] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 960.182547] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 960.182754] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 960.182910] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 960.183148] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 960.183344] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 960.183608] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 960.183803] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 960.183980] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 960.184193] env[63197]: DEBUG nova.virt.hardware [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 960.185197] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9684393c-4aaf-42da-9899-08d23b2a6281 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.194342] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251c1e20-ce1e-42fe-a952-4139eb87f35c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.210381] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.210736] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.210994] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.211230] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.211430] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.214019] env[63197]: INFO nova.compute.manager [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Terminating instance [ 960.216155] env[63197]: DEBUG nova.compute.manager [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.216460] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.217271] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3b966a-3d94-41cf-b6a2-2b4d3b27d62b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.222513] env[63197]: DEBUG nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 960.225838] env[63197]: DEBUG nova.network.neutron [-] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.230275] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.234161] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38f00b23-c02b-441d-9bde-3ae5a1234b7b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.242273] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364398, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.244127] env[63197]: DEBUG oslo_vmware.api [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 960.244127] env[63197]: value = "task-1364400" [ 960.244127] env[63197]: _type = "Task" [ 960.244127] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.254805] env[63197]: DEBUG oslo_vmware.api [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.266903] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364399, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.275252] env[63197]: DEBUG oslo_concurrency.lockutils [req-4a584bf6-4c4d-4917-b119-e59dccc69690 req-334c5bc3-59ce-4d82-8f9a-d232eb4367d5 service nova] Releasing lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.291807] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "c41b34cb-cc39-461e-83df-cd4de6780d95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.292101] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.292358] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "c41b34cb-cc39-461e-83df-cd4de6780d95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.292564] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.292742] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.295462] env[63197]: INFO nova.compute.manager [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Terminating instance [ 960.298189] env[63197]: DEBUG nova.compute.manager [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.298424] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.299345] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec4b408-bc77-4bfa-9964-6e3eb1d63bf2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.309891] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.310252] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a643c935-5a20-4419-b050-545b54ecd75e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.317927] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 960.317927] env[63197]: value = "task-1364401" [ 960.317927] env[63197]: _type = "Task" [ 960.317927] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.327334] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.414794] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.415402] env[63197]: DEBUG nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.638571] env[63197]: INFO nova.compute.manager [-] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Took 1.93 seconds to deallocate network for instance. [ 960.733082] env[63197]: INFO nova.compute.manager [-] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Took 1.61 seconds to deallocate network for instance. [ 960.749024] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364398, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60872} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.750346] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 8e5ef1b0-7532-498a-84c2-189274a36c50/8e5ef1b0-7532-498a-84c2-189274a36c50.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.750675] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.751773] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f77bc0f-50da-400a-851a-abb35225c82d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.759530] env[63197]: DEBUG oslo_vmware.api [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364400, 'name': PowerOffVM_Task, 'duration_secs': 0.404028} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.764418] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.764697] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 960.765166] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 960.765166] env[63197]: value = "task-1364402" [ 960.765166] env[63197]: _type = "Task" [ 960.765166] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.766311] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.766648] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.768232] env[63197]: INFO nova.compute.claims [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.772431] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8654d9eb-abce-4581-baae-3d9ff865bce7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.785301] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364399, 'name': RemoveSnapshot_Task, 'duration_secs': 0.902779} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.785408] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 960.785773] env[63197]: DEBUG nova.compute.manager [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.786944] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acc4386-7328-4a71-ba82-2fc1d8802b85 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.793566] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364402, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.830832] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364401, 'name': PowerOffVM_Task, 'duration_secs': 0.365601} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.831191] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.831421] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 960.831685] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-534e5fed-3b30-4b98-86b7-6bcb85320b22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.844781] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 960.845030] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 960.845227] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Deleting the datastore file [datastore2] 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.845495] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa82cd81-a110-4356-b645-962c46c7c2ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.853187] env[63197]: DEBUG oslo_vmware.api [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for the task: (returnval){ [ 960.853187] env[63197]: value = "task-1364405" [ 960.853187] env[63197]: _type = "Task" [ 960.853187] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.861469] env[63197]: DEBUG oslo_vmware.api [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.899637] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 960.899637] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 960.899637] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleting the datastore file [datastore1] c41b34cb-cc39-461e-83df-cd4de6780d95 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.899637] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3466e8a-0397-4a8f-a351-6d0c65b9b7a8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.908175] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 960.908175] env[63197]: value = "task-1364406" [ 960.908175] env[63197]: _type = "Task" [ 960.908175] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.918996] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364406, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.921658] env[63197]: DEBUG nova.compute.utils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.922115] env[63197]: DEBUG nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.922115] env[63197]: DEBUG nova.network.neutron [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.005042] env[63197]: DEBUG nova.policy [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4651bf2d103498cab54fb45a665277b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '61bfcd8038f046789676dc740a9aa991', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.142427] env[63197]: DEBUG nova.network.neutron [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Successfully updated port: 6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.146609] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.151090] env[63197]: DEBUG nova.compute.manager [req-f518b2e1-41a2-4620-b110-33258853f0a3 req-59f4a560-f3c8-442e-86a2-a491a9fd5e01 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-vif-plugged-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.152127] env[63197]: DEBUG oslo_concurrency.lockutils [req-f518b2e1-41a2-4620-b110-33258853f0a3 req-59f4a560-f3c8-442e-86a2-a491a9fd5e01 service nova] Acquiring lock "4672f595-e1f6-4400-b5a1-065598584980-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.152566] env[63197]: DEBUG oslo_concurrency.lockutils [req-f518b2e1-41a2-4620-b110-33258853f0a3 req-59f4a560-f3c8-442e-86a2-a491a9fd5e01 service nova] Lock "4672f595-e1f6-4400-b5a1-065598584980-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.152753] env[63197]: DEBUG oslo_concurrency.lockutils [req-f518b2e1-41a2-4620-b110-33258853f0a3 req-59f4a560-f3c8-442e-86a2-a491a9fd5e01 service nova] Lock "4672f595-e1f6-4400-b5a1-065598584980-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.152978] env[63197]: DEBUG nova.compute.manager [req-f518b2e1-41a2-4620-b110-33258853f0a3 req-59f4a560-f3c8-442e-86a2-a491a9fd5e01 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] No waiting events found dispatching network-vif-plugged-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.153184] env[63197]: WARNING nova.compute.manager [req-f518b2e1-41a2-4620-b110-33258853f0a3 req-59f4a560-f3c8-442e-86a2-a491a9fd5e01 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received unexpected event network-vif-plugged-6085d1e1-acf6-4471-8a56-e050285a562d for instance with vm_state building and task_state spawning. [ 961.250366] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.288709] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364402, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110851} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.288958] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.290343] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e24d7fe-cd7a-46dc-b027-5f8b752f018e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.320475] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 8e5ef1b0-7532-498a-84c2-189274a36c50/8e5ef1b0-7532-498a-84c2-189274a36c50.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.321086] env[63197]: INFO nova.compute.manager [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Shelve offloading [ 961.322887] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc9cf363-dd05-4d27-b019-361d8b66b224 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.341341] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.345021] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba2cf128-d893-47f7-bb94-559d08e099e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.349495] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 961.349495] env[63197]: value = "task-1364408" [ 961.349495] env[63197]: _type = "Task" [ 961.349495] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.351303] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 961.351303] env[63197]: value = "task-1364407" [ 961.351303] env[63197]: _type = "Task" [ 961.351303] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.365016] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364408, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.372664] env[63197]: DEBUG oslo_vmware.api [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Task: {'id': task-1364405, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29172} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.373049] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.373254] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.373443] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.373624] env[63197]: INFO nova.compute.manager [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Took 1.16 seconds to destroy the instance on the hypervisor. [ 961.373955] env[63197]: DEBUG oslo.service.loopingcall [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.374230] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] VM already powered off {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 961.374396] env[63197]: DEBUG nova.compute.manager [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.374659] env[63197]: DEBUG nova.compute.manager [-] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.374759] env[63197]: DEBUG nova.network.neutron [-] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 961.377177] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22038b2-276b-425c-8983-83105c8f458a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.383555] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.383741] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.383872] env[63197]: DEBUG nova.network.neutron [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.418899] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364406, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.425394] env[63197]: DEBUG nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.552462] env[63197]: DEBUG nova.network.neutron [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Successfully created port: f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.622117] env[63197]: DEBUG nova.compute.manager [req-56b799b1-472d-4009-9644-110deb6760f4 req-849bc0fb-701d-4188-8bdb-b97c39411e77 service nova] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Received event network-vif-deleted-720e1317-71ba-4079-a3fb-794d854e0e13 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.622512] env[63197]: DEBUG nova.compute.manager [req-56b799b1-472d-4009-9644-110deb6760f4 req-849bc0fb-701d-4188-8bdb-b97c39411e77 service nova] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Received event network-vif-deleted-2d53dd8b-1ba1-495d-99fc-265d17cb34a5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.645923] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.646170] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.646250] env[63197]: DEBUG nova.network.neutron [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 961.861060] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364408, 'name': ReconfigVM_Task, 'duration_secs': 0.346175} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.863816] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 8e5ef1b0-7532-498a-84c2-189274a36c50/8e5ef1b0-7532-498a-84c2-189274a36c50.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.864737] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0071dbb-6eed-4248-b8a6-2cc47026d426 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.871925] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 961.871925] env[63197]: value = "task-1364409" [ 961.871925] env[63197]: _type = "Task" [ 961.871925] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.883435] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364409, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.917940] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364406, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.018345] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c81927-3063-4aaf-82c3-fc5f6b524eec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.027162] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b6d073-77bb-4413-8fb7-5a1310a77b11 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.062469] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb52e0d-47b0-43f8-b18e-263b752b83ff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.074077] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acad791-ea79-42b5-8af5-1e06c03480db {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.085920] env[63197]: DEBUG nova.compute.provider_tree [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.143885] env[63197]: DEBUG nova.network.neutron [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62e2288c-e6", "ovs_interfaceid": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.192470] env[63197]: DEBUG nova.network.neutron [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 962.326415] env[63197]: DEBUG nova.network.neutron [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.383134] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364409, 'name': Rename_Task, 'duration_secs': 0.140651} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.383531] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.383886] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d68a48b7-d9cc-4686-9102-b1af8e297007 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.390995] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 962.390995] env[63197]: value = "task-1364410" [ 962.390995] env[63197]: _type = "Task" [ 962.390995] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.399621] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.417411] env[63197]: DEBUG oslo_vmware.api [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364406, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.101907} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.417745] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.417990] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.418205] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.418390] env[63197]: INFO nova.compute.manager [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Took 2.12 seconds to destroy the instance on the hypervisor. [ 962.418744] env[63197]: DEBUG oslo.service.loopingcall [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.418826] env[63197]: DEBUG nova.compute.manager [-] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.418956] env[63197]: DEBUG nova.network.neutron [-] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.436298] env[63197]: DEBUG nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.456880] env[63197]: DEBUG nova.network.neutron [-] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.460136] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.460420] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.460585] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.460777] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.460958] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.461155] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.461386] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.461552] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.461722] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.461889] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.462166] env[63197]: DEBUG nova.virt.hardware [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.462984] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92623791-3a4c-4331-9e58-85d42c2611bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.473020] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafd16ca-96a0-4b34-9cbd-3a1d105440aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.589112] env[63197]: DEBUG nova.scheduler.client.report [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.648759] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.813912] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c806bb-7df4-23d9-5d07-969dc9945cc5/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 962.817118] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa4e082-d6a6-4b50-8e66-d45fac0b3977 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.822297] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c806bb-7df4-23d9-5d07-969dc9945cc5/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 962.824086] env[63197]: ERROR oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c806bb-7df4-23d9-5d07-969dc9945cc5/disk-0.vmdk due to incomplete transfer. [ 962.824086] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-dafe1231-2a0b-40c5-8a2f-4c3a2a4e7492 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.834232] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.834232] env[63197]: DEBUG nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Instance network_info: |[{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.834232] env[63197]: DEBUG oslo_vmware.rw_handles [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c806bb-7df4-23d9-5d07-969dc9945cc5/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 962.834232] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Uploaded image 0b7db7fb-059d-491c-979a-4ff6763069f5 to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 962.834232] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 962.837046] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:3a:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6085d1e1-acf6-4471-8a56-e050285a562d', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.842552] env[63197]: DEBUG oslo.service.loopingcall [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.842804] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e30807e3-6367-4519-9f46-288265b9ef48 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.844594] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 962.846654] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e539cdf-02b1-498f-b823-9def69e0e93d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.864994] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 962.864994] env[63197]: value = "task-1364411" [ 962.864994] env[63197]: _type = "Task" [ 962.864994] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.866429] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.866429] env[63197]: value = "task-1364412" [ 962.866429] env[63197]: _type = "Task" [ 962.866429] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.880780] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364411, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.882606] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364412, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.901577] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364410, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.971317] env[63197]: INFO nova.compute.manager [-] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Took 1.59 seconds to deallocate network for instance. [ 963.100588] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.100588] env[63197]: DEBUG nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.101189] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.955s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.101608] env[63197]: DEBUG nova.objects.instance [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lazy-loading 'resources' on Instance uuid c97f59db-b1bf-4744-98ed-b6d7c095207f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.146220] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.146220] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33e1357-ed55-4ff8-b06b-7434b7e6e25e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.151128] env[63197]: DEBUG nova.network.neutron [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Successfully updated port: f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.160907] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 963.161545] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e418c40d-d75e-4a32-a5c7-6fc4e1e09b2a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.186731] env[63197]: DEBUG nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.187021] env[63197]: DEBUG nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing instance network info cache due to event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.187258] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.187412] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.187582] env[63197]: DEBUG nova.network.neutron [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.227977] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 963.228266] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 963.228437] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleting the datastore file [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 963.228882] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-436df4ef-26ef-4e48-8918-4dcf2112f2d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.234860] env[63197]: DEBUG nova.network.neutron [-] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.237207] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 963.237207] env[63197]: value = "task-1364414" [ 963.237207] env[63197]: _type = "Task" [ 963.237207] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.246703] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364414, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.381225] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364411, 'name': Destroy_Task, 'duration_secs': 0.41265} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.384340] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Destroyed the VM [ 963.384590] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 963.384958] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364412, 'name': CreateVM_Task, 'duration_secs': 0.437415} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.385162] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-823f7e93-9996-43f4-a05a-9b8fcecb9c45 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.386596] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 963.387347] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.387541] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.387856] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.388389] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc2739e4-801e-455a-a8ae-f89d2627735a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.391753] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 963.391753] env[63197]: value = "task-1364415" [ 963.391753] env[63197]: _type = "Task" [ 963.391753] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.396277] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 963.396277] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5206c32a-7992-839b-8a3d-d7ed2da99e42" [ 963.396277] env[63197]: _type = "Task" [ 963.396277] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.405350] env[63197]: DEBUG oslo_vmware.api [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364410, 'name': PowerOnVM_Task, 'duration_secs': 0.547272} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.405907] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.406138] env[63197]: INFO nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Took 8.28 seconds to spawn the instance on the hypervisor. [ 963.406340] env[63197]: DEBUG nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.409672] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a7a350-aaee-4a67-b185-48b214213602 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.411999] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364415, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.415303] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5206c32a-7992-839b-8a3d-d7ed2da99e42, 'name': SearchDatastore_Task, 'duration_secs': 0.009984} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.415872] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.416143] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.416386] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.416539] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.416719] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.416997] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25331dd9-175a-4864-8f1d-2113c8b0ed3a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.427588] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.427786] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 963.428479] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22bc6749-47d9-422d-a186-e09c231d691b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.432850] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 963.432850] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526c41af-cacc-9e59-e9ee-42828c92d1fe" [ 963.432850] env[63197]: _type = "Task" [ 963.432850] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.439814] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526c41af-cacc-9e59-e9ee-42828c92d1fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.481139] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.605899] env[63197]: DEBUG nova.compute.utils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.613016] env[63197]: DEBUG nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.613330] env[63197]: DEBUG nova.network.neutron [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 963.655768] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.655916] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.656086] env[63197]: DEBUG nova.network.neutron [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 963.670133] env[63197]: DEBUG nova.policy [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b50be4b75a94b4481c9c65ea1e4e9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bb89fb32d8c4726a9a3104d68ce560a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.687018] env[63197]: DEBUG nova.compute.manager [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Received event network-vif-deleted-e33a7c35-f308-47b5-8b23-08328d08d9e5 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.687248] env[63197]: DEBUG nova.compute.manager [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Received event network-vif-plugged-f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.687448] env[63197]: DEBUG oslo_concurrency.lockutils [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] Acquiring lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.687652] env[63197]: DEBUG oslo_concurrency.lockutils [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.687827] env[63197]: DEBUG oslo_concurrency.lockutils [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.687993] env[63197]: DEBUG nova.compute.manager [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] No waiting events found dispatching network-vif-plugged-f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.688181] env[63197]: WARNING nova.compute.manager [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Received unexpected event network-vif-plugged-f725a825-fe1f-4b06-b1df-63c5ac1b3295 for instance with vm_state building and task_state spawning. [ 963.688939] env[63197]: DEBUG nova.compute.manager [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Received event network-changed-f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.688939] env[63197]: DEBUG nova.compute.manager [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Refreshing instance network info cache due to event network-changed-f725a825-fe1f-4b06-b1df-63c5ac1b3295. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 963.688939] env[63197]: DEBUG oslo_concurrency.lockutils [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] Acquiring lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.740072] env[63197]: INFO nova.compute.manager [-] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Took 1.32 seconds to deallocate network for instance. [ 963.757736] env[63197]: DEBUG oslo_vmware.api [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364414, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138883} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.758029] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.758222] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 963.758407] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 963.781902] env[63197]: INFO nova.scheduler.client.report [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted allocations for instance 180e0da7-f7ee-4fcd-be95-c2bf679278d3 [ 963.832077] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a60f3db-af47-478b-8df2-3fe8ac6ee9cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.847503] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056dbf12-92b2-4dc7-be1c-7962a7425d25 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.888624] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f012dc90-2125-4f52-bd30-80d9beee360d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.901073] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfe8eb6-5a53-40af-b666-98f7a0715562 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.911951] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364415, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.920145] env[63197]: DEBUG nova.compute.provider_tree [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.932826] env[63197]: INFO nova.compute.manager [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Took 15.10 seconds to build instance. [ 963.945308] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526c41af-cacc-9e59-e9ee-42828c92d1fe, 'name': SearchDatastore_Task, 'duration_secs': 0.007626} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.945308] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68b3e5e7-9769-43e4-898c-94ba5c12beab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.949883] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 963.949883] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5206417b-09b9-2ce6-4761-339d720c84fb" [ 963.949883] env[63197]: _type = "Task" [ 963.949883] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.963806] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5206417b-09b9-2ce6-4761-339d720c84fb, 'name': SearchDatastore_Task, 'duration_secs': 0.009356} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.963806] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.963806] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 4672f595-e1f6-4400-b5a1-065598584980/4672f595-e1f6-4400-b5a1-065598584980.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 963.963806] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce5bab34-0554-4536-a1de-5b5b52f537e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.969011] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 963.969011] env[63197]: value = "task-1364416" [ 963.969011] env[63197]: _type = "Task" [ 963.969011] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.976628] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364416, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.033051] env[63197]: DEBUG nova.network.neutron [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Successfully created port: cbc85003-5324-451a-b9ba-25b9581079c7 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.114233] env[63197]: DEBUG nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.170350] env[63197]: DEBUG nova.network.neutron [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updated VIF entry in instance network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 964.172300] env[63197]: DEBUG nova.network.neutron [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.235212] env[63197]: DEBUG nova.network.neutron [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.252123] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.287028] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.411180] env[63197]: DEBUG oslo_vmware.api [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364415, 'name': RemoveSnapshot_Task, 'duration_secs': 0.98397} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.411483] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 964.411727] env[63197]: INFO nova.compute.manager [None req-bde102fd-4141-42d9-ae8c-abce680de1fc tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Took 14.69 seconds to snapshot the instance on the hypervisor. [ 964.424720] env[63197]: DEBUG nova.scheduler.client.report [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.438683] env[63197]: DEBUG oslo_concurrency.lockutils [None req-358d1b36-e471-4947-9645-779fad1ec884 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.622s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.482619] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364416, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.518746] env[63197]: DEBUG nova.network.neutron [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updating instance_info_cache with network_info: [{"id": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "address": "fa:16:3e:44:f4:a8", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf725a825-fe", "ovs_interfaceid": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.677023] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.677023] env[63197]: DEBUG nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Received event network-vif-deleted-ae6d2ab9-dd07-41b4-8a35-7c7c49b68bd9 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.677023] env[63197]: DEBUG nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-vif-unplugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.677023] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.677023] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.677023] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.677023] env[63197]: DEBUG nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] No waiting events found dispatching network-vif-unplugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.677023] env[63197]: WARNING nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received unexpected event network-vif-unplugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 for instance with vm_state shelved and task_state shelving_offloading. [ 964.677023] env[63197]: DEBUG nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.677023] env[63197]: DEBUG nova.compute.manager [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing instance network info cache due to event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 964.677023] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Acquiring lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.677023] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Acquired lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.677023] env[63197]: DEBUG nova.network.neutron [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 964.929688] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.931999] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.682s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.932428] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.934151] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.453s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.934381] env[63197]: DEBUG nova.objects.instance [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lazy-loading 'resources' on Instance uuid 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.961973] env[63197]: INFO nova.scheduler.client.report [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Deleted allocations for instance c97f59db-b1bf-4744-98ed-b6d7c095207f [ 964.963718] env[63197]: INFO nova.scheduler.client.report [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted allocations for instance 537cc8cc-94e0-41b4-8204-6f995f2ea4d3 [ 964.979865] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364416, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608407} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.980150] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 4672f595-e1f6-4400-b5a1-065598584980/4672f595-e1f6-4400-b5a1-065598584980.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 964.980379] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.980641] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d92a5e0-cdc0-489f-b775-2b5ce78095fa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.986859] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 964.986859] env[63197]: value = "task-1364417" [ 964.986859] env[63197]: _type = "Task" [ 964.986859] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.996074] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364417, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.022684] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Releasing lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.023019] env[63197]: DEBUG nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Instance network_info: |[{"id": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "address": "fa:16:3e:44:f4:a8", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf725a825-fe", "ovs_interfaceid": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.023372] env[63197]: DEBUG oslo_concurrency.lockutils [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] Acquired lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.023562] env[63197]: DEBUG nova.network.neutron [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Refreshing network info cache for port f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.025020] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:f4:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e30245c5-78f5-48e6-b504-c6c21f5a9b45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f725a825-fe1f-4b06-b1df-63c5ac1b3295', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.034014] env[63197]: DEBUG oslo.service.loopingcall [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.034014] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.034155] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-821541d6-3b73-4730-b457-2ec31e669f1f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.056549] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.056549] env[63197]: value = "task-1364418" [ 965.056549] env[63197]: _type = "Task" [ 965.056549] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.065555] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364418, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.127266] env[63197]: DEBUG nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.159313] env[63197]: DEBUG nova.virt.hardware [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.160088] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19735a07-757c-46f1-8c99-80d4ca065080 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.168757] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02d5bdc-eb77-4c85-9839-6d674ffd99c3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.215254] env[63197]: DEBUG nova.compute.manager [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Received event network-changed-fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.215529] env[63197]: DEBUG nova.compute.manager [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Refreshing instance network info cache due to event network-changed-fa39c111-5e72-43e8-b764-d2072c0c3d40. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.215682] env[63197]: DEBUG oslo_concurrency.lockutils [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] Acquiring lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.215855] env[63197]: DEBUG oslo_concurrency.lockutils [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] Acquired lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.215969] env[63197]: DEBUG nova.network.neutron [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Refreshing network info cache for port fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.418995] env[63197]: DEBUG nova.network.neutron [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updated VIF entry in instance network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.419405] env[63197]: DEBUG nova.network.neutron [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap62e2288c-e6", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.473350] env[63197]: DEBUG oslo_concurrency.lockutils [None req-af426455-e982-40ab-b0bd-703a6698506b tempest-ServersTestMultiNic-309248509 tempest-ServersTestMultiNic-309248509-project-member] Lock "c97f59db-b1bf-4744-98ed-b6d7c095207f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.955s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.474432] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ffd3765a-d3bb-4dba-b54e-ec47ff3e4b69 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "537cc8cc-94e0-41b4-8204-6f995f2ea4d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.479s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.504234] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364417, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068533} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.504544] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.505860] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d090b64-0380-4357-b0fe-d2547f2a910b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.531210] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 4672f595-e1f6-4400-b5a1-065598584980/4672f595-e1f6-4400-b5a1-065598584980.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.534126] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc33c3f8-2289-43f8-bfe4-ca3b49e62b9d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.562064] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 965.562064] env[63197]: value = "task-1364419" [ 965.562064] env[63197]: _type = "Task" [ 965.562064] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.568730] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364418, 'name': CreateVM_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.574799] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364419, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.666303] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb23967-361c-4e1b-8608-13b0b09fc568 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.674436] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56bfd6e3-2227-4ced-9825-a02ec01e1f8b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.716233] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f7bc31-c74d-4cbe-9493-8568ea0ba4ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.734941] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decd7f65-6e0e-4530-9467-8099a03ef46e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.743967] env[63197]: DEBUG nova.compute.manager [req-9d1dea60-7f78-4e42-a4e2-0c11ee4ba969 req-59816edd-12b2-4129-b8a0-f0b8e10ea40a service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Received event network-vif-plugged-cbc85003-5324-451a-b9ba-25b9581079c7 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.744344] env[63197]: DEBUG oslo_concurrency.lockutils [req-9d1dea60-7f78-4e42-a4e2-0c11ee4ba969 req-59816edd-12b2-4129-b8a0-f0b8e10ea40a service nova] Acquiring lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.744646] env[63197]: DEBUG oslo_concurrency.lockutils [req-9d1dea60-7f78-4e42-a4e2-0c11ee4ba969 req-59816edd-12b2-4129-b8a0-f0b8e10ea40a service nova] Lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.744887] env[63197]: DEBUG oslo_concurrency.lockutils [req-9d1dea60-7f78-4e42-a4e2-0c11ee4ba969 req-59816edd-12b2-4129-b8a0-f0b8e10ea40a service nova] Lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.745124] env[63197]: DEBUG nova.compute.manager [req-9d1dea60-7f78-4e42-a4e2-0c11ee4ba969 req-59816edd-12b2-4129-b8a0-f0b8e10ea40a service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] No waiting events found dispatching network-vif-plugged-cbc85003-5324-451a-b9ba-25b9581079c7 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.745444] env[63197]: WARNING nova.compute.manager [req-9d1dea60-7f78-4e42-a4e2-0c11ee4ba969 req-59816edd-12b2-4129-b8a0-f0b8e10ea40a service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Received unexpected event network-vif-plugged-cbc85003-5324-451a-b9ba-25b9581079c7 for instance with vm_state building and task_state spawning. [ 965.758136] env[63197]: DEBUG nova.compute.provider_tree [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.922215] env[63197]: DEBUG oslo_concurrency.lockutils [req-6741a272-0b8a-4083-8f51-d54fac340703 req-64301696-6aef-488e-b1c7-c49883bf84a6 service nova] Releasing lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.950732] env[63197]: DEBUG nova.network.neutron [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Successfully updated port: cbc85003-5324-451a-b9ba-25b9581079c7 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.956947] env[63197]: DEBUG nova.network.neutron [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updated VIF entry in instance network info cache for port f725a825-fe1f-4b06-b1df-63c5ac1b3295. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.957677] env[63197]: DEBUG nova.network.neutron [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updating instance_info_cache with network_info: [{"id": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "address": "fa:16:3e:44:f4:a8", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf725a825-fe", "ovs_interfaceid": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.069990] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364418, 'name': CreateVM_Task, 'duration_secs': 0.733452} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.071528] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.072401] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.072574] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.072919] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.076153] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ea49f1d-ac52-4ee2-849d-186540086749 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.078051] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364419, 'name': ReconfigVM_Task, 'duration_secs': 0.270626} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.078343] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 4672f595-e1f6-4400-b5a1-065598584980/4672f595-e1f6-4400-b5a1-065598584980.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.079316] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4caf1632-482d-41b6-bcb1-5f99c5ab3515 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.082204] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 966.082204] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fb2841-0afd-0315-dd47-0da9cf9a4364" [ 966.082204] env[63197]: _type = "Task" [ 966.082204] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.088323] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 966.088323] env[63197]: value = "task-1364420" [ 966.088323] env[63197]: _type = "Task" [ 966.088323] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.095500] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fb2841-0afd-0315-dd47-0da9cf9a4364, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.101042] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364420, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.215650] env[63197]: DEBUG nova.network.neutron [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Updated VIF entry in instance network info cache for port fa39c111-5e72-43e8-b764-d2072c0c3d40. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.216166] env[63197]: DEBUG nova.network.neutron [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Updating instance_info_cache with network_info: [{"id": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "address": "fa:16:3e:86:e6:7f", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa39c111-5e", "ovs_interfaceid": "fa39c111-5e72-43e8-b764-d2072c0c3d40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.228055] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.260727] env[63197]: DEBUG nova.scheduler.client.report [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.453271] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.453435] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.453743] env[63197]: DEBUG nova.network.neutron [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.460170] env[63197]: DEBUG oslo_concurrency.lockutils [req-7094b80e-3ec7-4b46-9f07-9be69af26500 req-b15160bb-943a-44fe-811b-1cb476d09288 service nova] Releasing lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.593266] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fb2841-0afd-0315-dd47-0da9cf9a4364, 'name': SearchDatastore_Task, 'duration_secs': 0.014031} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.596602] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.596846] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.597096] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.597249] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.597433] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.597707] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5e7f215-cf9f-4cd0-8285-8f388b8c1e7f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.604990] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364420, 'name': Rename_Task, 'duration_secs': 0.129013} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.606047] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.606315] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.606480] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.607156] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7cff8637-247c-42f0-a324-3e279ce053f5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.608538] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ae62911-8115-4994-9d6e-deabb3414e0c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.613361] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 966.613361] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5252d887-9516-34c2-807c-807620afbdbd" [ 966.613361] env[63197]: _type = "Task" [ 966.613361] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.620357] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 966.620357] env[63197]: value = "task-1364421" [ 966.620357] env[63197]: _type = "Task" [ 966.620357] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.624450] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5252d887-9516-34c2-807c-807620afbdbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.629188] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364421, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.722085] env[63197]: DEBUG oslo_concurrency.lockutils [req-225e05a8-d137-4f3c-b088-62d7d69a863e req-785732af-6ab9-41d7-8d95-f57987f4c3c0 service nova] Releasing lock "refresh_cache-8e5ef1b0-7532-498a-84c2-189274a36c50" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.769718] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.772625] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.520s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.772881] env[63197]: DEBUG nova.objects.instance [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lazy-loading 'resources' on Instance uuid c41b34cb-cc39-461e-83df-cd4de6780d95 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.801649] env[63197]: INFO nova.scheduler.client.report [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Deleted allocations for instance 5b9c6a1c-4212-4941-aa6c-364fb82f5e64 [ 966.984967] env[63197]: DEBUG nova.network.neutron [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 967.123619] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5252d887-9516-34c2-807c-807620afbdbd, 'name': SearchDatastore_Task, 'duration_secs': 0.008819} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.124501] env[63197]: DEBUG nova.network.neutron [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance_info_cache with network_info: [{"id": "cbc85003-5324-451a-b9ba-25b9581079c7", "address": "fa:16:3e:21:a3:85", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc85003-53", "ovs_interfaceid": "cbc85003-5324-451a-b9ba-25b9581079c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.129677] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27c64e9b-41b4-4a6c-826f-8891355c153c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.136997] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364421, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.138720] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 967.138720] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f68ab-87e3-3903-cd6b-7cc98f7864dc" [ 967.138720] env[63197]: _type = "Task" [ 967.138720] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.146812] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f68ab-87e3-3903-cd6b-7cc98f7864dc, 'name': SearchDatastore_Task, 'duration_secs': 0.009267} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.147062] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.147319] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373/10976c1e-e6ff-4db9-9ef2-a0ff3cc73373.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.147556] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3e3ca4f-3892-40c8-b380-aa0b0e520bb6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.153502] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 967.153502] env[63197]: value = "task-1364422" [ 967.153502] env[63197]: _type = "Task" [ 967.153502] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.161092] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364422, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.179678] env[63197]: DEBUG nova.compute.manager [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.180489] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b113570f-47ba-4989-b801-ad44cc15c040 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.310520] env[63197]: DEBUG oslo_concurrency.lockutils [None req-1c023aaf-2f18-4824-b675-7e44d62df839 tempest-AttachVolumeTestJSON-2138809928 tempest-AttachVolumeTestJSON-2138809928-project-member] Lock "5b9c6a1c-4212-4941-aa6c-364fb82f5e64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.100s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.439269] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cd6b48-f6dd-46e3-9954-b53a5690865a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.448110] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a31bfb2-b513-4e53-b5db-04d4c633be4b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.480660] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c4ad68-ff9e-4216-8aab-58e274c822cd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.488788] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5ef30d-d64e-43b8-82d1-7fca2a2065bc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.503173] env[63197]: DEBUG nova.compute.provider_tree [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.634036] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.634036] env[63197]: DEBUG nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Instance network_info: |[{"id": "cbc85003-5324-451a-b9ba-25b9581079c7", "address": "fa:16:3e:21:a3:85", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc85003-53", "ovs_interfaceid": "cbc85003-5324-451a-b9ba-25b9581079c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 967.635304] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:a3:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cbc85003-5324-451a-b9ba-25b9581079c7', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.644492] env[63197]: DEBUG oslo.service.loopingcall [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.645075] env[63197]: DEBUG oslo_vmware.api [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364421, 'name': PowerOnVM_Task, 'duration_secs': 0.674958} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.645800] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.646217] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 967.646597] env[63197]: INFO nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Took 7.49 seconds to spawn the instance on the hypervisor. [ 967.646922] env[63197]: DEBUG nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.647317] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-483c3c22-ca6f-40c6-a765-7127cd64a463 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.668789] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22454efb-87a5-429f-a2f9-4fc92fb1b5f0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.686826] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364422, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.690831] env[63197]: INFO nova.compute.manager [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] instance snapshotting [ 967.696615] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.696615] env[63197]: value = "task-1364423" [ 967.696615] env[63197]: _type = "Task" [ 967.696615] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.697988] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2bf74e-8b6d-454d-90b6-884ffb6b2cd8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.726940] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8424d65-3d35-4535-a543-7d097c7cdfe7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.729931] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364423, 'name': CreateVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.776726] env[63197]: DEBUG nova.compute.manager [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Received event network-changed-cbc85003-5324-451a-b9ba-25b9581079c7 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.776955] env[63197]: DEBUG nova.compute.manager [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Refreshing instance network info cache due to event network-changed-cbc85003-5324-451a-b9ba-25b9581079c7. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.777160] env[63197]: DEBUG oslo_concurrency.lockutils [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] Acquiring lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.777308] env[63197]: DEBUG oslo_concurrency.lockutils [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] Acquired lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.777482] env[63197]: DEBUG nova.network.neutron [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Refreshing network info cache for port cbc85003-5324-451a-b9ba-25b9581079c7 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.007050] env[63197]: DEBUG nova.scheduler.client.report [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.184790] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364422, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.612656} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.185106] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373/10976c1e-e6ff-4db9-9ef2-a0ff3cc73373.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.185373] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.185593] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83716994-c5d8-47a5-8bbc-566f2a4c68aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.192459] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 968.192459] env[63197]: value = "task-1364425" [ 968.192459] env[63197]: _type = "Task" [ 968.192459] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.206496] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.209818] env[63197]: INFO nova.compute.manager [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Took 14.00 seconds to build instance. [ 968.215407] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364423, 'name': CreateVM_Task, 'duration_secs': 0.333316} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.216040] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.216239] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.216404] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.216722] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.217024] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1d64426-c5e6-4f35-abb9-b09767a9bf8a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.221646] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 968.221646] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523354c3-7325-bd55-a497-a66a566d33a0" [ 968.221646] env[63197]: _type = "Task" [ 968.221646] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.229212] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523354c3-7325-bd55-a497-a66a566d33a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.240999] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 968.240999] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b95df9a9-c31e-42c0-99ec-e13a7e41bc4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.250701] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 968.250701] env[63197]: value = "task-1364426" [ 968.250701] env[63197]: _type = "Task" [ 968.250701] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.260990] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364426, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.511545] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.739s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.517532] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.230s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.517532] env[63197]: DEBUG nova.objects.instance [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'resources' on Instance uuid 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.547328] env[63197]: INFO nova.scheduler.client.report [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted allocations for instance c41b34cb-cc39-461e-83df-cd4de6780d95 [ 968.676796] env[63197]: DEBUG nova.network.neutron [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updated VIF entry in instance network info cache for port cbc85003-5324-451a-b9ba-25b9581079c7. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 968.677228] env[63197]: DEBUG nova.network.neutron [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance_info_cache with network_info: [{"id": "cbc85003-5324-451a-b9ba-25b9581079c7", "address": "fa:16:3e:21:a3:85", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc85003-53", "ovs_interfaceid": "cbc85003-5324-451a-b9ba-25b9581079c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.702805] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072942} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.703105] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.703983] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e79a71-4ae2-48da-be49-3aaa7e8272d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.719199] env[63197]: DEBUG oslo_concurrency.lockutils [None req-79ef994e-2766-4f62-abad-54401c626adc tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "4672f595-e1f6-4400-b5a1-065598584980" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.516s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.729148] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373/10976c1e-e6ff-4db9-9ef2-a0ff3cc73373.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.729898] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b518a8d8-47f0-4a44-94b1-3acf56b6bd63 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.754087] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523354c3-7325-bd55-a497-a66a566d33a0, 'name': SearchDatastore_Task, 'duration_secs': 0.009341} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.758084] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.758338] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.758577] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.758731] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.758909] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.759253] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 968.759253] env[63197]: value = "task-1364427" [ 968.759253] env[63197]: _type = "Task" [ 968.759253] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.759494] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26817f58-5514-453a-875a-43a37de2454e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.767619] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364426, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.772999] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364427, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.775608] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.775759] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.776535] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-700ce446-56a4-4293-98a6-7c591117ffa6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.782323] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 968.782323] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5215a2c6-814c-d99d-5d1e-bfba8fa26d40" [ 968.782323] env[63197]: _type = "Task" [ 968.782323] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.793695] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5215a2c6-814c-d99d-5d1e-bfba8fa26d40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.020008] env[63197]: DEBUG nova.objects.instance [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'numa_topology' on Instance uuid 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.057198] env[63197]: DEBUG oslo_concurrency.lockutils [None req-03854843-a2d0-4e7e-b981-c6325550d832 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "c41b34cb-cc39-461e-83df-cd4de6780d95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.765s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.180213] env[63197]: DEBUG oslo_concurrency.lockutils [req-a91e0588-2f1d-4a2f-86fd-5ed07a12796e req-c2bd9986-c7a6-4927-aa0d-9e652dd54f63 service nova] Releasing lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.272462] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364426, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.277676] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364427, 'name': ReconfigVM_Task, 'duration_secs': 0.506377} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.278314] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373/10976c1e-e6ff-4db9-9ef2-a0ff3cc73373.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.279266] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd4b0aa7-f050-46c9-8815-11078232b35b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.287493] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 969.287493] env[63197]: value = "task-1364428" [ 969.287493] env[63197]: _type = "Task" [ 969.287493] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.295238] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5215a2c6-814c-d99d-5d1e-bfba8fa26d40, 'name': SearchDatastore_Task, 'duration_secs': 0.01034} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.296422] env[63197]: DEBUG nova.compute.manager [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.296650] env[63197]: DEBUG nova.compute.manager [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing instance network info cache due to event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 969.296895] env[63197]: DEBUG oslo_concurrency.lockutils [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.297099] env[63197]: DEBUG oslo_concurrency.lockutils [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.297308] env[63197]: DEBUG nova.network.neutron [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 969.299633] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d285bca8-595e-4fcf-b3ca-a29fa2bb405a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.305501] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364428, 'name': Rename_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.308868] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 969.308868] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c4b4a5-b042-5c1d-e0f8-2d171f899b3d" [ 969.308868] env[63197]: _type = "Task" [ 969.308868] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.317549] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c4b4a5-b042-5c1d-e0f8-2d171f899b3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.522770] env[63197]: DEBUG nova.objects.base [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Object Instance<180e0da7-f7ee-4fcd-be95-c2bf679278d3> lazy-loaded attributes: resources,numa_topology {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 969.538562] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "4757f91b-21f1-4a16-bda3-729b12ddf86c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.538562] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.538562] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "4757f91b-21f1-4a16-bda3-729b12ddf86c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.538562] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.538562] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.539464] env[63197]: INFO nova.compute.manager [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Terminating instance [ 969.542694] env[63197]: DEBUG nova.compute.manager [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.543732] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.544037] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd9028b-a776-4878-ad93-268e6d9156ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.551894] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.552403] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a6cf8a10-836a-4b2d-ae63-350f713bedaf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.562231] env[63197]: DEBUG oslo_vmware.api [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 969.562231] env[63197]: value = "task-1364429" [ 969.562231] env[63197]: _type = "Task" [ 969.562231] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.578018] env[63197]: DEBUG oslo_vmware.api [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.677412] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65aecc0-c0e4-4486-a872-68a8362e8114 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.685269] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5deeec2-51fc-4b03-a715-13313c777930 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.719087] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9d57ca-cc7c-49a9-959e-2118c968315b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.726963] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723b462d-b44f-4ed3-a1e1-22c53b338004 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.741253] env[63197]: DEBUG nova.compute.provider_tree [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.764154] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364426, 'name': CreateSnapshot_Task, 'duration_secs': 1.22379} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.764451] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 969.765258] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54406924-16b5-470e-b7fa-3710d7499e56 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.798353] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364428, 'name': Rename_Task, 'duration_secs': 0.191337} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.798653] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.798952] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8eccbeda-5d30-4aec-80b8-534155245a63 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.807753] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 969.807753] env[63197]: value = "task-1364430" [ 969.807753] env[63197]: _type = "Task" [ 969.807753] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.819271] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364430, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.823406] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c4b4a5-b042-5c1d-e0f8-2d171f899b3d, 'name': SearchDatastore_Task, 'duration_secs': 0.009586} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.823713] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.824014] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 64b944ea-2397-45eb-a0ed-9f57263db998/64b944ea-2397-45eb-a0ed-9f57263db998.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.824598] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a5b2236-dd84-476c-b97d-c47298b981ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.830964] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 969.830964] env[63197]: value = "task-1364431" [ 969.830964] env[63197]: _type = "Task" [ 969.830964] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.840640] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.052075] env[63197]: DEBUG nova.network.neutron [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updated VIF entry in instance network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.052752] env[63197]: DEBUG nova.network.neutron [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.075152] env[63197]: DEBUG oslo_vmware.api [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364429, 'name': PowerOffVM_Task, 'duration_secs': 0.250718} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.075479] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.075682] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.075976] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-817e3858-40a0-4493-9191-5774affdf67b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.142701] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.142962] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.143332] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleting the datastore file [datastore2] 4757f91b-21f1-4a16-bda3-729b12ddf86c {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.143650] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-998022b2-071b-44b4-b577-2903d26aedc4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.156505] env[63197]: DEBUG oslo_vmware.api [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 970.156505] env[63197]: value = "task-1364433" [ 970.156505] env[63197]: _type = "Task" [ 970.156505] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.166802] env[63197]: DEBUG oslo_vmware.api [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364433, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.232948] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.233390] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.245086] env[63197]: DEBUG nova.scheduler.client.report [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.285186] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 970.287486] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9507314f-bd6a-423d-ac10-26885b22c0f3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.298259] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 970.298259] env[63197]: value = "task-1364434" [ 970.298259] env[63197]: _type = "Task" [ 970.298259] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.310391] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364434, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.322257] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364430, 'name': PowerOnVM_Task} progress is 74%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.340203] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364431, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.559054] env[63197]: DEBUG oslo_concurrency.lockutils [req-58240709-258f-41d2-a3e9-7da579ce2360 req-9aac35f9-14d1-41ea-a0f7-806a8dfbdff9 service nova] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.667084] env[63197]: DEBUG oslo_vmware.api [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364433, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260044} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.667428] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.667625] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.667810] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.667983] env[63197]: INFO nova.compute.manager [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 970.668254] env[63197]: DEBUG oslo.service.loopingcall [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.668471] env[63197]: DEBUG nova.compute.manager [-] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.668592] env[63197]: DEBUG nova.network.neutron [-] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.736637] env[63197]: DEBUG nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.750952] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.234s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.809155] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364434, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.819699] env[63197]: DEBUG oslo_vmware.api [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364430, 'name': PowerOnVM_Task, 'duration_secs': 0.917309} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.820292] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.820701] env[63197]: INFO nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Took 8.38 seconds to spawn the instance on the hypervisor. [ 970.821026] env[63197]: DEBUG nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.821898] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9834aab7-9ef6-4d12-bf82-166529a9b7ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.840497] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522041} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.840951] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 64b944ea-2397-45eb-a0ed-9f57263db998/64b944ea-2397-45eb-a0ed-9f57263db998.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.841326] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.841943] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f60bf054-1696-4e26-a8c5-a540880de051 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.851018] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 970.851018] env[63197]: value = "task-1364435" [ 970.851018] env[63197]: _type = "Task" [ 970.851018] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.858980] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364435, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.259181] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fe3453e9-b591-4f57-8d79-bbdf3c3eac43 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.391s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.260393] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 5.034s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.260638] env[63197]: INFO nova.compute.manager [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Unshelving [ 971.265194] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.265440] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.267311] env[63197]: INFO nova.compute.claims [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.310472] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364434, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.318803] env[63197]: DEBUG nova.compute.manager [req-36bc5694-f3cf-4d1c-80ba-4a5d52df8b73 req-66052516-9efe-4864-81db-c72b0d909107 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Received event network-vif-deleted-4bebe341-10f9-4ff8-a0b7-3312eb20074f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.319039] env[63197]: INFO nova.compute.manager [req-36bc5694-f3cf-4d1c-80ba-4a5d52df8b73 req-66052516-9efe-4864-81db-c72b0d909107 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Neutron deleted interface 4bebe341-10f9-4ff8-a0b7-3312eb20074f; detaching it from the instance and deleting it from the info cache [ 971.319681] env[63197]: DEBUG nova.network.neutron [req-36bc5694-f3cf-4d1c-80ba-4a5d52df8b73 req-66052516-9efe-4864-81db-c72b0d909107 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.339950] env[63197]: INFO nova.compute.manager [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Took 15.64 seconds to build instance. [ 971.359166] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364435, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073004} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.359413] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.360133] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5381c4f-07ad-497a-bdf6-40028482c5dc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.381707] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 64b944ea-2397-45eb-a0ed-9f57263db998/64b944ea-2397-45eb-a0ed-9f57263db998.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.381979] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-beffdf24-f777-4396-b322-748fdc8be4e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.401138] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 971.401138] env[63197]: value = "task-1364437" [ 971.401138] env[63197]: _type = "Task" [ 971.401138] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.408710] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364437, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.596487] env[63197]: DEBUG nova.network.neutron [-] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.809959] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364434, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.822214] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-614de936-7cea-4bfe-98a2-ed5b70d407ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.832339] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6848c08f-f706-4265-9ae3-9b72170161d0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.843592] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e00fbb45-b360-4f8e-ba43-1dea673a351b tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.151s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.859839] env[63197]: DEBUG nova.compute.manager [req-36bc5694-f3cf-4d1c-80ba-4a5d52df8b73 req-66052516-9efe-4864-81db-c72b0d909107 service nova] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Detach interface failed, port_id=4bebe341-10f9-4ff8-a0b7-3312eb20074f, reason: Instance 4757f91b-21f1-4a16-bda3-729b12ddf86c could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 971.911792] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364437, 'name': ReconfigVM_Task, 'duration_secs': 0.274779} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.912189] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 64b944ea-2397-45eb-a0ed-9f57263db998/64b944ea-2397-45eb-a0ed-9f57263db998.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.913051] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e1fb8e8-96c6-4858-bec2-2ddd87094056 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.920621] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 971.920621] env[63197]: value = "task-1364438" [ 971.920621] env[63197]: _type = "Task" [ 971.920621] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.937071] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364438, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.100030] env[63197]: INFO nova.compute.manager [-] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Took 1.43 seconds to deallocate network for instance. [ 972.269897] env[63197]: DEBUG nova.compute.utils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.308704] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364434, 'name': CloneVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.408512] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613417f6-3467-42fb-9ded-32a6ccb00d3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.416883] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140ffb2a-d106-475a-bef7-45969ba22e0f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.451820] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36edbe90-a7b6-40f3-9e29-33ceb5b0bd77 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.455268] env[63197]: DEBUG nova.compute.manager [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Received event network-changed-f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 972.455460] env[63197]: DEBUG nova.compute.manager [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Refreshing instance network info cache due to event network-changed-f725a825-fe1f-4b06-b1df-63c5ac1b3295. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 972.455744] env[63197]: DEBUG oslo_concurrency.lockutils [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] Acquiring lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.455900] env[63197]: DEBUG oslo_concurrency.lockutils [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] Acquired lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.456078] env[63197]: DEBUG nova.network.neutron [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Refreshing network info cache for port f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.461149] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364438, 'name': Rename_Task, 'duration_secs': 0.346019} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.461718] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.461958] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66694497-3d55-43ab-810f-60e9e592b305 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.467682] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8520324d-77f9-4980-9526-fc90d84479a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.472931] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 972.472931] env[63197]: value = "task-1364439" [ 972.472931] env[63197]: _type = "Task" [ 972.472931] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.485941] env[63197]: DEBUG nova.compute.provider_tree [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.495215] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364439, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.608096] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.772994] env[63197]: INFO nova.virt.block_device [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Booting with volume 652f45ea-384f-474b-9fd4-23d569d7d473 at /dev/sdb [ 972.805998] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3cd731aa-ad78-4855-9034-badcda30425d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.813695] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364434, 'name': CloneVM_Task, 'duration_secs': 2.127136} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.814951] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Created linked-clone VM from snapshot [ 972.815715] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d339dd-6d8e-43fd-99ad-c14062b63667 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.820754] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c67b46-eb06-4f78-99be-0581a58591de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.850705] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bddf2676-e7ec-455d-a171-1f0678f1364a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.852554] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Uploading image 5e0d7e86-5084-413c-ac64-8430a5c0417b {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 972.860489] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d50045-dcbd-4a14-96df-ea9ce4cea3b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.879876] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 972.879876] env[63197]: value = "vm-290382" [ 972.879876] env[63197]: _type = "VirtualMachine" [ 972.879876] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 972.880219] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e1c8f0a4-c810-4768-9575-8135818de990 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.892166] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25e82f5-e878-43b8-ba48-2c4a1893867d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.894999] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lease: (returnval){ [ 972.894999] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277ac06-e943-eb07-e1d8-08f32f8ba1f1" [ 972.894999] env[63197]: _type = "HttpNfcLease" [ 972.894999] env[63197]: } obtained for exporting VM: (result){ [ 972.894999] env[63197]: value = "vm-290382" [ 972.894999] env[63197]: _type = "VirtualMachine" [ 972.894999] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 972.895364] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the lease: (returnval){ [ 972.895364] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277ac06-e943-eb07-e1d8-08f32f8ba1f1" [ 972.895364] env[63197]: _type = "HttpNfcLease" [ 972.895364] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 972.901206] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9831ba54-9844-482a-9e14-3fb7f53b0550 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.904594] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 972.904594] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277ac06-e943-eb07-e1d8-08f32f8ba1f1" [ 972.904594] env[63197]: _type = "HttpNfcLease" [ 972.904594] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 972.915128] env[63197]: DEBUG nova.virt.block_device [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating existing volume attachment record: 62e695c2-8e3c-40a8-b76b-c0c8c6f6c20e {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 972.984859] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364439, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.988801] env[63197]: DEBUG nova.scheduler.client.report [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.039061] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.039446] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.039697] env[63197]: INFO nova.compute.manager [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Rebooting instance [ 973.192096] env[63197]: DEBUG nova.network.neutron [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updated VIF entry in instance network info cache for port f725a825-fe1f-4b06-b1df-63c5ac1b3295. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 973.192709] env[63197]: DEBUG nova.network.neutron [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updating instance_info_cache with network_info: [{"id": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "address": "fa:16:3e:44:f4:a8", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf725a825-fe", "ovs_interfaceid": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.403562] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 973.403562] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277ac06-e943-eb07-e1d8-08f32f8ba1f1" [ 973.403562] env[63197]: _type = "HttpNfcLease" [ 973.403562] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 973.403886] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 973.403886] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277ac06-e943-eb07-e1d8-08f32f8ba1f1" [ 973.403886] env[63197]: _type = "HttpNfcLease" [ 973.403886] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 973.404645] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6464c1e4-6bef-4327-8b58-a0d827a76775 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.411848] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bf5c8a-fa07-a805-383f-ab9b30c9b701/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 973.412062] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bf5c8a-fa07-a805-383f-ab9b30c9b701/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 973.484748] env[63197]: DEBUG oslo_vmware.api [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364439, 'name': PowerOnVM_Task, 'duration_secs': 0.742427} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.485244] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.485593] env[63197]: INFO nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Took 8.36 seconds to spawn the instance on the hypervisor. [ 973.485892] env[63197]: DEBUG nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.486786] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d000a8-b2d8-40b5-a556-24ec3930f15c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.496126] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.496606] env[63197]: DEBUG nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.499309] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.891s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.499542] env[63197]: DEBUG nova.objects.instance [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lazy-loading 'resources' on Instance uuid 4757f91b-21f1-4a16-bda3-729b12ddf86c {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.508968] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8787f4f4-4bb6-4ce3-8351-881145b27fdd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.567833] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.695625] env[63197]: DEBUG oslo_concurrency.lockutils [req-fa3b61bb-dab6-4a8e-904a-2f48887ab927 req-b5b72dcd-b1b6-40be-b111-527055b8c06a service nova] Releasing lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.696603] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquired lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.696981] env[63197]: DEBUG nova.network.neutron [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 974.006114] env[63197]: DEBUG nova.compute.utils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.012445] env[63197]: DEBUG nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.012857] env[63197]: DEBUG nova.network.neutron [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 974.014902] env[63197]: INFO nova.compute.manager [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Took 13.28 seconds to build instance. [ 974.083342] env[63197]: DEBUG nova.policy [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.190169] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e085d497-c46a-486d-b648-811a07cd8f54 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.198368] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf8e57cb-5ff0-44b3-84b2-9d54f1985aa4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.236943] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de10f14-cf56-439a-8df0-d73b2dbdaef4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.247544] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb32c9d-271d-4bf3-93ce-35e2ce6d1992 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.260928] env[63197]: DEBUG nova.compute.provider_tree [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.469145] env[63197]: DEBUG nova.network.neutron [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Successfully created port: 6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.508525] env[63197]: DEBUG nova.network.neutron [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updating instance_info_cache with network_info: [{"id": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "address": "fa:16:3e:44:f4:a8", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf725a825-fe", "ovs_interfaceid": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.511182] env[63197]: DEBUG nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.517071] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5143f2d9-7e3c-4b00-afcf-eff673a22424 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.803s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.767319] env[63197]: DEBUG nova.scheduler.client.report [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 975.016131] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Releasing lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.026929] env[63197]: DEBUG nova.compute.manager [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.028275] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b2b278-003d-4667-bf48-f834da538ab7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.273153] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.774s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.294294] env[63197]: INFO nova.scheduler.client.report [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted allocations for instance 4757f91b-21f1-4a16-bda3-729b12ddf86c [ 975.537032] env[63197]: DEBUG nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.583588] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.584905] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.584905] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.585059] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.585177] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.586356] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.586637] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.586868] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.587172] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.587461] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.587740] env[63197]: DEBUG nova.virt.hardware [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.588735] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83bdfce-e038-4a53-9827-f45b50189c3a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.597855] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7399722c-7aa0-46f3-84d6-ffed341d14ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.811033] env[63197]: DEBUG oslo_concurrency.lockutils [None req-44d35830-bb5f-4773-85da-a298c7c4a0c3 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "4757f91b-21f1-4a16-bda3-729b12ddf86c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.273s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.054842] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335c2683-e11b-4a8a-94ef-1673f01a100b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.064076] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Doing hard reboot of VM {{(pid=63197) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 976.064476] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-239931b1-ef44-4e9d-8fdf-7d7b0c5e2e66 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.072055] env[63197]: DEBUG oslo_vmware.api [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 976.072055] env[63197]: value = "task-1364445" [ 976.072055] env[63197]: _type = "Task" [ 976.072055] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.084252] env[63197]: DEBUG oslo_vmware.api [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364445, 'name': ResetVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.177551] env[63197]: DEBUG nova.compute.manager [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Stashing vm_state: active {{(pid=63197) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 976.331876] env[63197]: DEBUG nova.compute.manager [req-078ac0c2-b131-43ff-8c83-ca5ab82ef172 req-ef6d134f-03d3-4f48-89bd-b80aee1663b8 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-vif-plugged-6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.332099] env[63197]: DEBUG oslo_concurrency.lockutils [req-078ac0c2-b131-43ff-8c83-ca5ab82ef172 req-ef6d134f-03d3-4f48-89bd-b80aee1663b8 service nova] Acquiring lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.332347] env[63197]: DEBUG oslo_concurrency.lockutils [req-078ac0c2-b131-43ff-8c83-ca5ab82ef172 req-ef6d134f-03d3-4f48-89bd-b80aee1663b8 service nova] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.332609] env[63197]: DEBUG oslo_concurrency.lockutils [req-078ac0c2-b131-43ff-8c83-ca5ab82ef172 req-ef6d134f-03d3-4f48-89bd-b80aee1663b8 service nova] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.332813] env[63197]: DEBUG nova.compute.manager [req-078ac0c2-b131-43ff-8c83-ca5ab82ef172 req-ef6d134f-03d3-4f48-89bd-b80aee1663b8 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] No waiting events found dispatching network-vif-plugged-6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.333053] env[63197]: WARNING nova.compute.manager [req-078ac0c2-b131-43ff-8c83-ca5ab82ef172 req-ef6d134f-03d3-4f48-89bd-b80aee1663b8 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received unexpected event network-vif-plugged-6de5e286-38ff-452d-b567-9b6c956831d1 for instance with vm_state building and task_state spawning. [ 976.581981] env[63197]: DEBUG oslo_vmware.api [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364445, 'name': ResetVM_Task, 'duration_secs': 0.101123} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.585778] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Did hard reboot of VM {{(pid=63197) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 976.585778] env[63197]: DEBUG nova.compute.manager [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.585778] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d1da1f-201e-4f67-b77d-56e11f862010 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.700106] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.700206] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.754548] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.754627] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.106916] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e96914fb-2459-4c33-9297-b2aee342fe88 tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.067s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.169341] env[63197]: DEBUG nova.network.neutron [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Successfully updated port: 6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.193326] env[63197]: DEBUG nova.compute.manager [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.193659] env[63197]: DEBUG nova.compute.manager [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing instance network info cache due to event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.193861] env[63197]: DEBUG oslo_concurrency.lockutils [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.194037] env[63197]: DEBUG oslo_concurrency.lockutils [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.194176] env[63197]: DEBUG nova.network.neutron [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing network info cache for port 6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.206240] env[63197]: INFO nova.compute.claims [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.256600] env[63197]: DEBUG nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.603692] env[63197]: DEBUG nova.compute.manager [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Received event network-changed-f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.603984] env[63197]: DEBUG nova.compute.manager [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Refreshing instance network info cache due to event network-changed-f725a825-fe1f-4b06-b1df-63c5ac1b3295. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.604234] env[63197]: DEBUG oslo_concurrency.lockutils [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] Acquiring lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.604403] env[63197]: DEBUG oslo_concurrency.lockutils [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] Acquired lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.604579] env[63197]: DEBUG nova.network.neutron [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Refreshing network info cache for port f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.673218] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.713029] env[63197]: INFO nova.compute.resource_tracker [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating resource usage from migration 5d03443b-9ea8-4897-a373-50aca57af1de [ 977.727200] env[63197]: DEBUG nova.network.neutron [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.780571] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.817891] env[63197]: DEBUG nova.network.neutron [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.907053] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1d6e52-82bd-407f-927d-f8e6150d1fa1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.916164] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9fc9aa-62e2-4281-9edf-f186c7c0663d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.949415] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45fa660-f1da-41aa-b02b-345e0541608c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.957905] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e8a114-aa38-44ba-9d2f-e855ea6b2696 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.976013] env[63197]: DEBUG nova.compute.provider_tree [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.316564] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.316564] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.316564] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.316564] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.316564] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.316564] env[63197]: INFO nova.compute.manager [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Terminating instance [ 978.316564] env[63197]: DEBUG nova.compute.manager [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 978.316564] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.316564] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d474eba1-6279-4d06-a9ea-7a3c5b2251cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.316564] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.316564] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89f86018-ccf0-41aa-8d9e-814c02f69273 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.316564] env[63197]: DEBUG oslo_vmware.api [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 978.316564] env[63197]: value = "task-1364446" [ 978.316564] env[63197]: _type = "Task" [ 978.316564] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.316564] env[63197]: DEBUG oslo_vmware.api [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364446, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.320790] env[63197]: DEBUG oslo_concurrency.lockutils [req-41688bda-ffe2-4b1a-a708-0d7137759b48 req-3bf36efc-8e96-41dd-8dff-fdda02337e58 service nova] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.321285] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.321462] env[63197]: DEBUG nova.network.neutron [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.422537] env[63197]: DEBUG nova.network.neutron [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updated VIF entry in instance network info cache for port f725a825-fe1f-4b06-b1df-63c5ac1b3295. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.422999] env[63197]: DEBUG nova.network.neutron [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updating instance_info_cache with network_info: [{"id": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "address": "fa:16:3e:44:f4:a8", "network": {"id": "941db9a2-4450-483e-b02d-b49348bd4d8a", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-922993710-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "61bfcd8038f046789676dc740a9aa991", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e30245c5-78f5-48e6-b504-c6c21f5a9b45", "external-id": "nsx-vlan-transportzone-409", "segmentation_id": 409, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf725a825-fe", "ovs_interfaceid": "f725a825-fe1f-4b06-b1df-63c5ac1b3295", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.479412] env[63197]: DEBUG nova.scheduler.client.report [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.530897] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.781717] env[63197]: DEBUG oslo_vmware.api [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364446, 'name': PowerOffVM_Task, 'duration_secs': 0.339283} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.781998] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.782245] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.782485] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3c691ec-7124-4e4d-bba6-96e30f8377a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.859938] env[63197]: DEBUG nova.network.neutron [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 978.866189] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.866189] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.866326] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Deleting the datastore file [datastore2] 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.866608] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90b3db8a-8bab-4e61-be60-141cde36aaf6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.873014] env[63197]: DEBUG oslo_vmware.api [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 978.873014] env[63197]: value = "task-1364448" [ 978.873014] env[63197]: _type = "Task" [ 978.873014] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.882138] env[63197]: DEBUG oslo_vmware.api [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364448, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.926208] env[63197]: DEBUG oslo_concurrency.lockutils [req-03660615-a42b-45fa-a81c-dd3b27199831 req-0b7c2438-f1bd-414b-a3d0-308db8f1b900 service nova] Releasing lock "refresh_cache-10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.984964] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.284s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.984964] env[63197]: INFO nova.compute.manager [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Migrating [ 978.991677] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.211s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.993329] env[63197]: INFO nova.compute.claims [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.005400] env[63197]: DEBUG nova.network.neutron [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.385174] env[63197]: DEBUG oslo_vmware.api [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364448, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.418135} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.385658] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.385755] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.385978] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.386211] env[63197]: INFO nova.compute.manager [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Took 1.14 seconds to destroy the instance on the hypervisor. [ 979.386521] env[63197]: DEBUG oslo.service.loopingcall [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.386756] env[63197]: DEBUG nova.compute.manager [-] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.386877] env[63197]: DEBUG nova.network.neutron [-] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 979.504301] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.504521] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.504840] env[63197]: DEBUG nova.network.neutron [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.510086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.511551] env[63197]: DEBUG nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Instance network_info: |[{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 979.511551] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:7e:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '298bb8ef-4765-494c-b157-7a349218bd1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6de5e286-38ff-452d-b567-9b6c956831d1', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.523288] env[63197]: DEBUG oslo.service.loopingcall [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.525358] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 979.525579] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-930e618d-75ca-433e-b9e7-e081a89038bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.555047] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.555047] env[63197]: value = "task-1364449" [ 979.555047] env[63197]: _type = "Task" [ 979.555047] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.565210] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364449, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.679725] env[63197]: DEBUG nova.compute.manager [req-11c45391-64c7-4215-8d3b-f4303360afa4 req-5f8d653a-4214-4de0-9ef9-5071b2b33170 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Received event network-vif-deleted-f725a825-fe1f-4b06-b1df-63c5ac1b3295 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.679949] env[63197]: INFO nova.compute.manager [req-11c45391-64c7-4215-8d3b-f4303360afa4 req-5f8d653a-4214-4de0-9ef9-5071b2b33170 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Neutron deleted interface f725a825-fe1f-4b06-b1df-63c5ac1b3295; detaching it from the instance and deleting it from the info cache [ 979.680150] env[63197]: DEBUG nova.network.neutron [req-11c45391-64c7-4215-8d3b-f4303360afa4 req-5f8d653a-4214-4de0-9ef9-5071b2b33170 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.071069] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364449, 'name': CreateVM_Task, 'duration_secs': 0.434524} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.071069] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 980.071880] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.072127] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.072600] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 980.076259] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a7e01a0-0707-4f29-909d-0610c130e029 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.082590] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 980.082590] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c087b7-d407-6b67-7657-c08915bdb650" [ 980.082590] env[63197]: _type = "Task" [ 980.082590] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.090463] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c087b7-d407-6b67-7657-c08915bdb650, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.157210] env[63197]: DEBUG nova.network.neutron [-] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.184183] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55658639-c332-4b55-be29-e1b7cd5d2cc0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.192721] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901d9c21-f47b-4f04-8a67-9a7750aa7d30 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.202638] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70e169e-f7bc-4bcf-87b7-92912158217c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.209604] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3370e2-3cc5-4ac9-aea5-ab3023f7e203 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.260774] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e84719-d8fd-4651-a733-c9111b583220 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.263942] env[63197]: DEBUG nova.compute.manager [req-11c45391-64c7-4215-8d3b-f4303360afa4 req-5f8d653a-4214-4de0-9ef9-5071b2b33170 service nova] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Detach interface failed, port_id=f725a825-fe1f-4b06-b1df-63c5ac1b3295, reason: Instance 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.271459] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9892ac42-1f06-4348-8385-5ed654063296 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.285721] env[63197]: DEBUG nova.compute.provider_tree [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.310062] env[63197]: DEBUG nova.network.neutron [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance_info_cache with network_info: [{"id": "cbc85003-5324-451a-b9ba-25b9581079c7", "address": "fa:16:3e:21:a3:85", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc85003-53", "ovs_interfaceid": "cbc85003-5324-451a-b9ba-25b9581079c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.593865] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c087b7-d407-6b67-7657-c08915bdb650, 'name': SearchDatastore_Task, 'duration_secs': 0.014414} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.594276] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.594582] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.594868] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.595051] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.595299] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.595605] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2727c1ed-fd8d-4bec-a685-314758371196 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.604822] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.605223] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 980.605966] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bebbf88d-37e8-4c6f-9db3-5c635792410b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.611296] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 980.611296] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52905c32-6804-2e29-11ac-a906a201a2c8" [ 980.611296] env[63197]: _type = "Task" [ 980.611296] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.619853] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52905c32-6804-2e29-11ac-a906a201a2c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.659641] env[63197]: INFO nova.compute.manager [-] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Took 1.27 seconds to deallocate network for instance. [ 980.788209] env[63197]: DEBUG nova.scheduler.client.report [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.812658] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.123871] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52905c32-6804-2e29-11ac-a906a201a2c8, 'name': SearchDatastore_Task, 'duration_secs': 0.014699} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.124846] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-002d43f6-d50c-4145-b346-946f706354e3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.131289] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 981.131289] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525538c6-f910-439d-c7c8-78568db73bc1" [ 981.131289] env[63197]: _type = "Task" [ 981.131289] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.139964] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525538c6-f910-439d-c7c8-78568db73bc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.166299] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.293561] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.294121] env[63197]: DEBUG nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.296967] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.766s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.297284] env[63197]: DEBUG nova.objects.instance [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'pci_requests' on Instance uuid 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.641352] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]525538c6-f910-439d-c7c8-78568db73bc1, 'name': SearchDatastore_Task, 'duration_secs': 0.01928} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.641698] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.641880] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f/1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 981.642162] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e11a1057-93d5-446c-ac78-7e5a219460f9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.648517] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 981.648517] env[63197]: value = "task-1364450" [ 981.648517] env[63197]: _type = "Task" [ 981.648517] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.656030] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364450, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.801606] env[63197]: DEBUG nova.compute.utils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.804405] env[63197]: DEBUG nova.objects.instance [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'numa_topology' on Instance uuid 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.806863] env[63197]: DEBUG nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.807781] env[63197]: DEBUG nova.network.neutron [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 981.865438] env[63197]: DEBUG nova.policy [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12d00044b65c48f484ac9fea595d6179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80138785cfdb4a2188ffb972da437646', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.128834] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.128834] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.128834] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 982.161583] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364450, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.238493] env[63197]: DEBUG nova.network.neutron [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Successfully created port: ed419ac7-cefa-4394-ad05-c3c85c6111e3 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 982.314220] env[63197]: DEBUG nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.314913] env[63197]: INFO nova.compute.claims [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.330941] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab224d1-a45d-4026-90f3-58e0fd28c459 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.350091] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance '64b944ea-2397-45eb-a0ed-9f57263db998' progress to 0 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 982.660756] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364450, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53782} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.661044] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f/1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 982.661266] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.661522] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74adc9ee-edc5-4ba6-9584-ce84fad8b53a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.668301] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 982.668301] env[63197]: value = "task-1364451" [ 982.668301] env[63197]: _type = "Task" [ 982.668301] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.675929] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364451, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.856770] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.857115] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23393bf2-0284-4039-897f-6bc88abf7be2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.865126] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 982.865126] env[63197]: value = "task-1364452" [ 982.865126] env[63197]: _type = "Task" [ 982.865126] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.875503] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.180312] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364451, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066637} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.180312] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 983.181208] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6fda9fa-b771-46f5-88ff-de4ba5190671 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.204935] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f/1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.205654] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9053c67f-ddec-44af-8ab2-3c011ac10598 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.225885] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 983.225885] env[63197]: value = "task-1364453" [ 983.225885] env[63197]: _type = "Task" [ 983.225885] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.234516] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364453, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.324582] env[63197]: DEBUG nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.348035] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.348314] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.348548] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.348789] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.348946] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.349171] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.349442] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.349643] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.349851] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.350066] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.350297] env[63197]: DEBUG nova.virt.hardware [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.351182] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054ea378-f585-4c63-90a7-b9f30851d788 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.362825] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db3b38b-96ec-4d07-9e20-3fda22fe2643 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.385245] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364452, 'name': PowerOffVM_Task, 'duration_secs': 0.372079} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.387817] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.388057] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance '64b944ea-2397-45eb-a0ed-9f57263db998' progress to 17 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 983.393012] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bf5c8a-fa07-a805-383f-ab9b30c9b701/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 983.393950] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ca3720-6485-43c8-a048-5aeee0d181a7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.399796] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bf5c8a-fa07-a805-383f-ab9b30c9b701/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 983.399972] env[63197]: ERROR oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bf5c8a-fa07-a805-383f-ab9b30c9b701/disk-0.vmdk due to incomplete transfer. [ 983.402304] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-def503bb-e0c0-4d77-b147-158bf8f56447 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.409203] env[63197]: DEBUG oslo_vmware.rw_handles [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52bf5c8a-fa07-a805-383f-ab9b30c9b701/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 983.409411] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Uploaded image 5e0d7e86-5084-413c-ac64-8430a5c0417b to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 983.411706] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 983.411960] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1fb8a214-ca04-4dd3-b5cd-b39af7879db0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.419986] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 983.419986] env[63197]: value = "task-1364454" [ 983.419986] env[63197]: _type = "Task" [ 983.419986] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.429684] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364454, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.492964] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73d571f-02c0-4214-9aec-f4ec4b4e1adf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.501055] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be07f73-580d-4679-9eed-bb464260934f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.532811] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088db942-dd6f-4839-9992-145dbf476d27 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.540431] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed012f31-a907-4e98-b5f1-7d28ed9d68ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.553997] env[63197]: DEBUG nova.compute.provider_tree [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.670854] env[63197]: DEBUG nova.compute.manager [req-291126a2-ccc9-4977-8f24-57963664f329 req-9aa80074-d1cc-45c4-9884-5bbd2549333b service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Received event network-vif-plugged-ed419ac7-cefa-4394-ad05-c3c85c6111e3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.671222] env[63197]: DEBUG oslo_concurrency.lockutils [req-291126a2-ccc9-4977-8f24-57963664f329 req-9aa80074-d1cc-45c4-9884-5bbd2549333b service nova] Acquiring lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.671307] env[63197]: DEBUG oslo_concurrency.lockutils [req-291126a2-ccc9-4977-8f24-57963664f329 req-9aa80074-d1cc-45c4-9884-5bbd2549333b service nova] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.672297] env[63197]: DEBUG oslo_concurrency.lockutils [req-291126a2-ccc9-4977-8f24-57963664f329 req-9aa80074-d1cc-45c4-9884-5bbd2549333b service nova] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.672297] env[63197]: DEBUG nova.compute.manager [req-291126a2-ccc9-4977-8f24-57963664f329 req-9aa80074-d1cc-45c4-9884-5bbd2549333b service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] No waiting events found dispatching network-vif-plugged-ed419ac7-cefa-4394-ad05-c3c85c6111e3 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.672297] env[63197]: WARNING nova.compute.manager [req-291126a2-ccc9-4977-8f24-57963664f329 req-9aa80074-d1cc-45c4-9884-5bbd2549333b service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Received unexpected event network-vif-plugged-ed419ac7-cefa-4394-ad05-c3c85c6111e3 for instance with vm_state building and task_state spawning. [ 983.738017] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.751121] env[63197]: DEBUG nova.network.neutron [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Successfully updated port: ed419ac7-cefa-4394-ad05-c3c85c6111e3 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.899870] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.899870] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.900257] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.900257] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.900401] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.900586] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.900797] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.900961] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.901145] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.901312] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.901511] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.906968] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-141332a3-d384-4d41-9f65-0b37b3042f38 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.921549] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 983.921549] env[63197]: value = "task-1364455" [ 983.921549] env[63197]: _type = "Task" [ 983.921549] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.932261] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364454, 'name': Destroy_Task, 'duration_secs': 0.487818} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.935178] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Destroyed the VM [ 983.935428] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 983.935711] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364455, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.936040] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a968f1d8-3df0-4f98-88bb-f1e9ef46cd1e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.941325] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 983.941325] env[63197]: value = "task-1364456" [ 983.941325] env[63197]: _type = "Task" [ 983.941325] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.949258] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364456, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.057112] env[63197]: DEBUG nova.scheduler.client.report [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.236388] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364453, 'name': ReconfigVM_Task, 'duration_secs': 0.61643} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.236682] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f/1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.237325] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9a87882-6266-4195-88db-0f2f2c0990a1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.243656] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 984.243656] env[63197]: value = "task-1364457" [ 984.243656] env[63197]: _type = "Task" [ 984.243656] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.253030] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364457, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.254703] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.254834] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.254998] env[63197]: DEBUG nova.network.neutron [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.436023] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364455, 'name': ReconfigVM_Task, 'duration_secs': 0.304299} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.436509] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance '64b944ea-2397-45eb-a0ed-9f57263db998' progress to 33 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 984.452786] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364456, 'name': RemoveSnapshot_Task} progress is 15%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.562352] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.265s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.564787] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.399s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.564940] env[63197]: DEBUG nova.objects.instance [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lazy-loading 'resources' on Instance uuid 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.609916] env[63197]: INFO nova.network.neutron [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 984.753907] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364457, 'name': Rename_Task, 'duration_secs': 0.242237} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.754160] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 984.754335] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa674706-f70c-4d42-b30c-e8828c4c90ba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.761373] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 984.761373] env[63197]: value = "task-1364458" [ 984.761373] env[63197]: _type = "Task" [ 984.761373] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.773482] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364458, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.788295] env[63197]: DEBUG nova.network.neutron [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 984.919088] env[63197]: DEBUG nova.network.neutron [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Updating instance_info_cache with network_info: [{"id": "ed419ac7-cefa-4394-ad05-c3c85c6111e3", "address": "fa:16:3e:76:e8:23", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped419ac7-ce", "ovs_interfaceid": "ed419ac7-cefa-4394-ad05-c3c85c6111e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.943225] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.943506] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.943736] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.943878] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.944048] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.944261] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.944508] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.944694] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.944875] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.945078] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.945286] env[63197]: DEBUG nova.virt.hardware [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.950934] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Reconfiguring VM instance instance-00000057 to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 984.951528] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf60ace3-3f3f-4259-bbce-dbff607d7d9a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.973670] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364456, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.974980] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 984.974980] env[63197]: value = "task-1364459" [ 984.974980] env[63197]: _type = "Task" [ 984.974980] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.982636] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364459, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.214985] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dd2d68-edcb-4d06-9d9e-7a5a0d150f9c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.224140] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc3f1cdd-5ff8-4496-b10a-1d96c990ec54 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.253758] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc536e9-de4b-4ef0-8c3b-aed35220d01a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.261610] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e290a0f-d32c-4256-8590-46fd8de79063 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.278186] env[63197]: DEBUG nova.compute.provider_tree [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.282591] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364458, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.422167] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.422511] env[63197]: DEBUG nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Instance network_info: |[{"id": "ed419ac7-cefa-4394-ad05-c3c85c6111e3", "address": "fa:16:3e:76:e8:23", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped419ac7-ce", "ovs_interfaceid": "ed419ac7-cefa-4394-ad05-c3c85c6111e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.422958] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:e8:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed419ac7-cefa-4394-ad05-c3c85c6111e3', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.432124] env[63197]: DEBUG oslo.service.loopingcall [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.432371] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.432626] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96cf9e57-2cec-47c7-96dd-88ad80851141 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.451676] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.451676] env[63197]: value = "task-1364460" [ 985.451676] env[63197]: _type = "Task" [ 985.451676] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.461541] env[63197]: DEBUG oslo_vmware.api [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364456, 'name': RemoveSnapshot_Task, 'duration_secs': 1.157579} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.464248] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 985.464485] env[63197]: INFO nova.compute.manager [None req-f609f94f-8f1a-48d9-b254-4a8300289175 tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Took 17.77 seconds to snapshot the instance on the hypervisor. [ 985.466792] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364460, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.483787] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364459, 'name': ReconfigVM_Task, 'duration_secs': 0.188854} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.484127] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Reconfigured VM instance instance-00000057 to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 985.484954] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f079dcb-1ab4-4f8c-a227-85491962a4cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.508636] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 64b944ea-2397-45eb-a0ed-9f57263db998/64b944ea-2397-45eb-a0ed-9f57263db998.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.509649] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-394a48e7-4f87-437c-9651-a166a0c310ea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.528727] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 985.528727] env[63197]: value = "task-1364461" [ 985.528727] env[63197]: _type = "Task" [ 985.528727] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.537075] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364461, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.700934] env[63197]: DEBUG nova.compute.manager [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Received event network-changed-ed419ac7-cefa-4394-ad05-c3c85c6111e3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.701208] env[63197]: DEBUG nova.compute.manager [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Refreshing instance network info cache due to event network-changed-ed419ac7-cefa-4394-ad05-c3c85c6111e3. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 985.701465] env[63197]: DEBUG oslo_concurrency.lockutils [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] Acquiring lock "refresh_cache-430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.701656] env[63197]: DEBUG oslo_concurrency.lockutils [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] Acquired lock "refresh_cache-430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.701874] env[63197]: DEBUG nova.network.neutron [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Refreshing network info cache for port ed419ac7-cefa-4394-ad05-c3c85c6111e3 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.776028] env[63197]: DEBUG oslo_vmware.api [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364458, 'name': PowerOnVM_Task, 'duration_secs': 0.524896} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.776487] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 985.776659] env[63197]: INFO nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Took 10.24 seconds to spawn the instance on the hypervisor. [ 985.776904] env[63197]: DEBUG nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 985.777746] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde95847-e72e-4507-9375-9ef2bb66d0b0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.784345] env[63197]: DEBUG nova.scheduler.client.report [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.964505] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364460, 'name': CreateVM_Task, 'duration_secs': 0.369736} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.964739] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.965504] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.965749] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.966150] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.966457] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57a4d1d3-ee71-425d-b5f4-062fc8722638 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.975443] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 985.975443] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eb999a-aa02-e406-142c-bdb279b04f77" [ 985.975443] env[63197]: _type = "Task" [ 985.975443] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.983436] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eb999a-aa02-e406-142c-bdb279b04f77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.038525] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364461, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.146356] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.146489] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.146685] env[63197]: DEBUG nova.network.neutron [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.151165] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Didn't find any instances for network info cache update. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 986.151387] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.151644] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.151853] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.152061] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.152266] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.152417] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.152692] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 986.152908] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.291845] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.303193] env[63197]: INFO nova.compute.manager [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Took 15.06 seconds to build instance. [ 986.314744] env[63197]: INFO nova.scheduler.client.report [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Deleted allocations for instance 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373 [ 986.421236] env[63197]: DEBUG nova.network.neutron [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Updated VIF entry in instance network info cache for port ed419ac7-cefa-4394-ad05-c3c85c6111e3. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.421689] env[63197]: DEBUG nova.network.neutron [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Updating instance_info_cache with network_info: [{"id": "ed419ac7-cefa-4394-ad05-c3c85c6111e3", "address": "fa:16:3e:76:e8:23", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped419ac7-ce", "ovs_interfaceid": "ed419ac7-cefa-4394-ad05-c3c85c6111e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.487062] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eb999a-aa02-e406-142c-bdb279b04f77, 'name': SearchDatastore_Task, 'duration_secs': 0.010356} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.487405] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.487661] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.487948] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.488068] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.488255] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.488550] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a0bf553-6fbc-4f1d-97c4-64991ab2d810 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.497376] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.497602] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.498332] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9edf1360-9a2f-4e4e-82c8-00cba50de5d6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.504018] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 986.504018] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5246b811-6d9a-abb6-d642-ece797cc968c" [ 986.504018] env[63197]: _type = "Task" [ 986.504018] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.512070] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5246b811-6d9a-abb6-d642-ece797cc968c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.538312] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364461, 'name': ReconfigVM_Task, 'duration_secs': 0.742478} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.538757] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 64b944ea-2397-45eb-a0ed-9f57263db998/64b944ea-2397-45eb-a0ed-9f57263db998.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.538995] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance '64b944ea-2397-45eb-a0ed-9f57263db998' progress to 50 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 986.656996] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.656996] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.656996] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.656996] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 986.657926] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edec3bf-9d1c-4902-9bd2-5630774a513a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.665757] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f559877-639c-4725-8a70-d8d1892fb858 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.679828] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2e6e1d-133c-4d66-9885-402bcf71085a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.687113] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5905b875-90bf-4f57-acc9-01ecfcab6bb2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.717999] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180285MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 986.718262] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.718350] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.805280] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4ab88f4f-de52-46cf-9778-8b2b72d8d948 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.572s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.824841] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5c373ec-bfa4-440f-a43e-d55c23324b3a tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "10976c1e-e6ff-4db9-9ef2-a0ff3cc73373" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.587s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.927334] env[63197]: DEBUG oslo_concurrency.lockutils [req-9572847f-e111-4afd-824e-33c104022a9b req-0c844951-d0d7-4c4e-8eb2-d0be561459f1 service nova] Releasing lock "refresh_cache-430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.014802] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5246b811-6d9a-abb6-d642-ece797cc968c, 'name': SearchDatastore_Task, 'duration_secs': 0.01331} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.015665] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce0d6559-e0a9-4dfa-bc4f-51d045e61835 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.023019] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 987.023019] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5258540f-ac58-29d6-f7a8-f06079b21c89" [ 987.023019] env[63197]: _type = "Task" [ 987.023019] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.028342] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5258540f-ac58-29d6-f7a8-f06079b21c89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.045699] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a59d80f-53d9-40e7-ad3e-315bf880f43a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.064614] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85761957-95af-480d-a50c-092b3761b88b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.082560] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance '64b944ea-2397-45eb-a0ed-9f57263db998' progress to 67 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 987.181132] env[63197]: DEBUG nova.network.neutron [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62e2288c-e6", "ovs_interfaceid": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.231507] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "4d12bac4-da42-42bd-9361-7015f3be9693" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.231833] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "4d12bac4-da42-42bd-9361-7015f3be9693" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.232056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "4d12bac4-da42-42bd-9361-7015f3be9693-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.232257] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "4d12bac4-da42-42bd-9361-7015f3be9693-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.232522] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "4d12bac4-da42-42bd-9361-7015f3be9693-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.234594] env[63197]: INFO nova.compute.manager [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Terminating instance [ 987.236801] env[63197]: DEBUG nova.compute.manager [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.237000] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.237828] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e8c8e8-6fef-400c-97fa-4d4c7be3e98d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.245677] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 987.246410] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94c29d4f-9a4a-42e7-a4d8-c0cdd405e46b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.253676] env[63197]: DEBUG oslo_vmware.api [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 987.253676] env[63197]: value = "task-1364462" [ 987.253676] env[63197]: _type = "Task" [ 987.253676] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.261224] env[63197]: DEBUG oslo_vmware.api [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364462, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.538019] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5258540f-ac58-29d6-f7a8-f06079b21c89, 'name': SearchDatastore_Task, 'duration_secs': 0.019724} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.538019] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.538019] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa/430a73fa-ae22-45dd-a5c7-8c5ed616ebaa.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.538019] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3515bbe9-acf7-48ca-a7d6-c7e0f347df5b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.543479] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 987.543479] env[63197]: value = "task-1364463" [ 987.543479] env[63197]: _type = "Task" [ 987.543479] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.553223] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364463, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.683459] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.715148] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c494f852568322a3c9a44d3215c721ab',container_format='bare',created_at=2024-10-25T09:50:13Z,direct_url=,disk_format='vmdk',id=073c5ea5-cd55-4c9a-ab31-9918df041ed1,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1255127347-shelved',owner='f35fa478269e4de68ae997e79237b25f',properties=ImageMetaProps,protected=,size=31665664,status='active',tags=,updated_at=2024-10-25T09:50:28Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.715444] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.715623] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.715880] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.716105] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.716388] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.718400] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.718400] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.718400] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.718400] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.718400] env[63197]: DEBUG nova.virt.hardware [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.718400] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3eaebf2-e040-4df5-8ade-19e0f43c4588 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.728220] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03cb7cd9-73b3-40bc-92df-3f3cdd84a0d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.734831] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Applying migration context for instance 64b944ea-2397-45eb-a0ed-9f57263db998 as it has an incoming, in-progress migration 5d03443b-9ea8-4897-a373-50aca57af1de. Migration status is migrating {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 987.736049] env[63197]: INFO nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating resource usage from migration 5d03443b-9ea8-4897-a373-50aca57af1de [ 987.753972] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:ba:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62e2288c-e61e-467a-bd6d-1ac6fd79cb57', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.767047] env[63197]: DEBUG oslo.service.loopingcall [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.768103] env[63197]: DEBUG nova.network.neutron [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Port cbc85003-5324-451a-b9ba-25b9581079c7 binding to destination host cpu-1 is already ACTIVE {{(pid=63197) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 987.770583] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.771978] env[63197]: DEBUG nova.compute.manager [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-vif-plugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.772227] env[63197]: DEBUG oslo_concurrency.lockutils [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.773043] env[63197]: DEBUG oslo_concurrency.lockutils [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.773043] env[63197]: DEBUG oslo_concurrency.lockutils [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.773043] env[63197]: DEBUG nova.compute.manager [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] No waiting events found dispatching network-vif-plugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.773250] env[63197]: WARNING nova.compute.manager [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received unexpected event network-vif-plugged-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 for instance with vm_state shelved_offloaded and task_state spawning. [ 987.773312] env[63197]: DEBUG nova.compute.manager [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.773473] env[63197]: DEBUG nova.compute.manager [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing instance network info cache due to event network-changed-62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.773683] env[63197]: DEBUG oslo_concurrency.lockutils [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] Acquiring lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.773885] env[63197]: DEBUG oslo_concurrency.lockutils [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] Acquired lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.774066] env[63197]: DEBUG nova.network.neutron [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Refreshing network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.779559] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f99dbfd-a1b2-4d7d-844e-699a3bc6f490 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.798314] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 638ef9c9-253b-4958-a660-6c1801408a51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.798532] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 4d12bac4-da42-42bd-9361-7015f3be9693 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.798626] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance c6686a09-3711-434b-b69f-a2518366fbed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.798755] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 8e5ef1b0-7532-498a-84c2-189274a36c50 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.798863] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 4672f595-e1f6-4400-b5a1-065598584980 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.799280] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 180e0da7-f7ee-4fcd-be95-c2bf679278d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.799439] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.799605] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Migration 5d03443b-9ea8-4897-a373-50aca57af1de is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 987.799833] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 64b944ea-2397-45eb-a0ed-9f57263db998 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.799979] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 987.800221] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 987.800471] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 987.814953] env[63197]: DEBUG oslo_vmware.api [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364462, 'name': PowerOffVM_Task, 'duration_secs': 0.195426} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.816808] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 987.817015] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.817290] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.817290] env[63197]: value = "task-1364464" [ 987.817290] env[63197]: _type = "Task" [ 987.817290] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.817479] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-beca8e5a-ab41-4229-a61c-1d0e24bf0a6c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.832118] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364464, 'name': CreateVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.918563] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.918881] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.919085] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Deleting the datastore file [datastore1] 4d12bac4-da42-42bd-9361-7015f3be9693 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.921752] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36258959-2361-49c3-bc0b-b658c0d11b42 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.928775] env[63197]: DEBUG oslo_vmware.api [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for the task: (returnval){ [ 987.928775] env[63197]: value = "task-1364466" [ 987.928775] env[63197]: _type = "Task" [ 987.928775] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.939591] env[63197]: DEBUG oslo_vmware.api [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364466, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.990324] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1a86da-999e-4c5d-8723-6f3c862cf4e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.999176] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60c4b90-de68-4382-a943-ec55ca3265aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.034682] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc908b68-4022-4010-8576-968735bba8c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.043490] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdaf2f3c-554a-4bc2-b212-0b018aa4a9cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.061309] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.065988] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364463, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.100562] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "c6686a09-3711-434b-b69f-a2518366fbed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.100851] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "c6686a09-3711-434b-b69f-a2518366fbed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.101086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "c6686a09-3711-434b-b69f-a2518366fbed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.101287] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "c6686a09-3711-434b-b69f-a2518366fbed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.101462] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "c6686a09-3711-434b-b69f-a2518366fbed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.106046] env[63197]: INFO nova.compute.manager [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Terminating instance [ 988.108138] env[63197]: DEBUG nova.compute.manager [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.108345] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.109217] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb59007-5629-497d-9729-62a156c1acc2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.117266] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.117720] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e8b2b60-5580-45b0-8e5c-f9f92a0cad35 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.124726] env[63197]: DEBUG oslo_vmware.api [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 988.124726] env[63197]: value = "task-1364467" [ 988.124726] env[63197]: _type = "Task" [ 988.124726] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.133136] env[63197]: DEBUG oslo_vmware.api [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.144810] env[63197]: DEBUG nova.network.neutron [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updated VIF entry in instance network info cache for port 62e2288c-e61e-467a-bd6d-1ac6fd79cb57. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.145288] env[63197]: DEBUG nova.network.neutron [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [{"id": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "address": "fa:16:3e:ba:ba:f7", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62e2288c-e6", "ovs_interfaceid": "62e2288c-e61e-467a-bd6d-1ac6fd79cb57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.330534] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364464, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.438143] env[63197]: DEBUG oslo_vmware.api [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Task: {'id': task-1364466, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257585} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.438414] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.439025] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 988.439025] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 988.439025] env[63197]: INFO nova.compute.manager [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Took 1.20 seconds to destroy the instance on the hypervisor. [ 988.439214] env[63197]: DEBUG oslo.service.loopingcall [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.439397] env[63197]: DEBUG nova.compute.manager [-] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.439496] env[63197]: DEBUG nova.network.neutron [-] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 988.557818] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364463, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.584277} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.558109] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa/430a73fa-ae22-45dd-a5c7-8c5ed616ebaa.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.558331] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.558584] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10945754-349f-44ef-a4ed-deb85222cb38 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.565230] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 988.565230] env[63197]: value = "task-1364468" [ 988.565230] env[63197]: _type = "Task" [ 988.565230] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.569539] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.580920] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.634734] env[63197]: DEBUG oslo_vmware.api [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364467, 'name': PowerOffVM_Task, 'duration_secs': 0.199091} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.634846] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.636070] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.636070] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e68ec8ba-83a3-431c-b652-2d50fed611cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.647694] env[63197]: DEBUG oslo_concurrency.lockutils [req-b0d8e81e-2608-46b4-9814-700cb7892113 req-a3a26615-6ffd-46b6-bdac-82dab2aa8940 service nova] Releasing lock "refresh_cache-180e0da7-f7ee-4fcd-be95-c2bf679278d3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.780703] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 988.781015] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 988.781220] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Deleting the datastore file [datastore1] c6686a09-3711-434b-b69f-a2518366fbed {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.781513] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c46ef0ca-be4b-4398-b936-1d773c106c71 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.803940] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.803940] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.803940] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.804473] env[63197]: DEBUG oslo_vmware.api [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for the task: (returnval){ [ 988.804473] env[63197]: value = "task-1364470" [ 988.804473] env[63197]: _type = "Task" [ 988.804473] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.815101] env[63197]: DEBUG oslo_vmware.api [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.829206] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364464, 'name': CreateVM_Task, 'duration_secs': 0.587055} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.829655] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 988.830462] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.830462] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.830654] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 988.830889] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb011b16-11cb-4c0a-8086-4da95e1db5db {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.836711] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 988.836711] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c1108b-4725-caa3-8477-143fedb98f97" [ 988.836711] env[63197]: _type = "Task" [ 988.836711] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.845325] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c1108b-4725-caa3-8477-143fedb98f97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.075136] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260683} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.075136] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.075885] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49eabd5-6d03-4380-b84c-3a7770ea3e61 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.078938] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 989.079136] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.361s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.100256] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa/430a73fa-ae22-45dd-a5c7-8c5ed616ebaa.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.100650] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd26867e-2d13-454f-9382-c2f0782299fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.121645] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 989.121645] env[63197]: value = "task-1364471" [ 989.121645] env[63197]: _type = "Task" [ 989.121645] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.130724] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364471, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.152281] env[63197]: DEBUG nova.network.neutron [-] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.267153] env[63197]: DEBUG nova.compute.manager [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.267481] env[63197]: DEBUG nova.compute.manager [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing instance network info cache due to event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.267598] env[63197]: DEBUG oslo_concurrency.lockutils [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.267758] env[63197]: DEBUG oslo_concurrency.lockutils [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.267920] env[63197]: DEBUG nova.network.neutron [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing network info cache for port 6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.314607] env[63197]: DEBUG oslo_vmware.api [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.346328] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.346558] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Processing image 073c5ea5-cd55-4c9a-ab31-9918df041ed1 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.346797] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1/073c5ea5-cd55-4c9a-ab31-9918df041ed1.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.346958] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1/073c5ea5-cd55-4c9a-ab31-9918df041ed1.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.347153] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.347403] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c0404e2-f990-4ffa-8d3b-7b06131c7f06 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.357018] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.357263] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.357936] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-254eb7e9-43bc-46c2-805b-c85f6f462b31 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.363565] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 989.363565] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e66b6e-7fb7-d9fd-827e-797301e3a55e" [ 989.363565] env[63197]: _type = "Task" [ 989.363565] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.371944] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e66b6e-7fb7-d9fd-827e-797301e3a55e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.630968] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364471, 'name': ReconfigVM_Task, 'duration_secs': 0.308693} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.631264] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa/430a73fa-ae22-45dd-a5c7-8c5ed616ebaa.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.631885] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a4fe7fa-167b-4140-887d-c338566265ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.638742] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 989.638742] env[63197]: value = "task-1364472" [ 989.638742] env[63197]: _type = "Task" [ 989.638742] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.646383] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364472, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.655053] env[63197]: INFO nova.compute.manager [-] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Took 1.22 seconds to deallocate network for instance. [ 989.793540] env[63197]: DEBUG nova.compute.manager [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.793822] env[63197]: DEBUG nova.compute.manager [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing instance network info cache due to event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.794087] env[63197]: DEBUG oslo_concurrency.lockutils [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.794269] env[63197]: DEBUG oslo_concurrency.lockutils [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.794444] env[63197]: DEBUG nova.network.neutron [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.817845] env[63197]: DEBUG oslo_vmware.api [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Task: {'id': task-1364470, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.572966} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.818186] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.818331] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 989.818460] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 989.818787] env[63197]: INFO nova.compute.manager [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Took 1.71 seconds to destroy the instance on the hypervisor. [ 989.818877] env[63197]: DEBUG oslo.service.loopingcall [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.819060] env[63197]: DEBUG nova.compute.manager [-] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.819485] env[63197]: DEBUG nova.network.neutron [-] [instance: c6686a09-3711-434b-b69f-a2518366fbed] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 989.853475] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.854574] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.854574] env[63197]: DEBUG nova.network.neutron [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.875060] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Preparing fetch location {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 989.875336] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Fetch image to [datastore2] OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1/OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1.vmdk {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 989.875526] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Downloading stream optimized image 073c5ea5-cd55-4c9a-ab31-9918df041ed1 to [datastore2] OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1/OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1.vmdk on the data store datastore2 as vApp {{(pid=63197) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 989.875703] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Downloading image file data 073c5ea5-cd55-4c9a-ab31-9918df041ed1 to the ESX as VM named 'OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1' {{(pid=63197) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 989.960597] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 989.960597] env[63197]: value = "resgroup-9" [ 989.960597] env[63197]: _type = "ResourcePool" [ 989.960597] env[63197]: }. {{(pid=63197) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 989.960902] env[63197]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ee9ceef9-37bd-4e72-8a80-ef800ae945ca {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.985175] env[63197]: DEBUG nova.network.neutron [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updated VIF entry in instance network info cache for port 6de5e286-38ff-452d-b567-9b6c956831d1. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 989.985533] env[63197]: DEBUG nova.network.neutron [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.991664] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease: (returnval){ [ 989.991664] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d4491-a46e-3ff1-cfe5-e475e89cc9d3" [ 989.991664] env[63197]: _type = "HttpNfcLease" [ 989.991664] env[63197]: } obtained for vApp import into resource pool (val){ [ 989.991664] env[63197]: value = "resgroup-9" [ 989.991664] env[63197]: _type = "ResourcePool" [ 989.991664] env[63197]: }. {{(pid=63197) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 989.991921] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the lease: (returnval){ [ 989.991921] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d4491-a46e-3ff1-cfe5-e475e89cc9d3" [ 989.991921] env[63197]: _type = "HttpNfcLease" [ 989.991921] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 990.001046] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.001046] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d4491-a46e-3ff1-cfe5-e475e89cc9d3" [ 990.001046] env[63197]: _type = "HttpNfcLease" [ 990.001046] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 990.149597] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364472, 'name': Rename_Task, 'duration_secs': 0.176488} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.149891] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.150188] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-61f96d1b-8f30-43b5-ba13-da6850946f1a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.156527] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 990.156527] env[63197]: value = "task-1364474" [ 990.156527] env[63197]: _type = "Task" [ 990.156527] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.164903] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.165184] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.165408] env[63197]: DEBUG nova.objects.instance [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lazy-loading 'resources' on Instance uuid 4d12bac4-da42-42bd-9361-7015f3be9693 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.166458] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.492949] env[63197]: DEBUG oslo_concurrency.lockutils [req-b725fc9a-adb3-497b-8a28-657bce90b7ea req-2a9eef93-99cf-40b8-9769-8ac7a8d217b4 service nova] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.500954] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.500954] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d4491-a46e-3ff1-cfe5-e475e89cc9d3" [ 990.500954] env[63197]: _type = "HttpNfcLease" [ 990.500954] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 990.558682] env[63197]: DEBUG nova.network.neutron [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance_info_cache with network_info: [{"id": "cbc85003-5324-451a-b9ba-25b9581079c7", "address": "fa:16:3e:21:a3:85", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc85003-53", "ovs_interfaceid": "cbc85003-5324-451a-b9ba-25b9581079c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.574644] env[63197]: DEBUG nova.network.neutron [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updated VIF entry in instance network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.575057] env[63197]: DEBUG nova.network.neutron [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.596146] env[63197]: DEBUG nova.network.neutron [-] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.666963] env[63197]: DEBUG oslo_vmware.api [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364474, 'name': PowerOnVM_Task, 'duration_secs': 0.459066} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.668062] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.668062] env[63197]: INFO nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Took 7.34 seconds to spawn the instance on the hypervisor. [ 990.668219] env[63197]: DEBUG nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.671218] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df86280-f49e-4370-9847-1052adc8ab27 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.807102] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad1e632-8a21-4dd5-9df3-815ff8e69e62 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.815943] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a2ec38-735c-4671-9fe4-cad9867748f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.848050] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e9836a-8864-4621-ab53-0ed9976f33c9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.855696] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebde2c58-dcb8-44d8-988d-d8349360dd92 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.870128] env[63197]: DEBUG nova.compute.provider_tree [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.002189] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 991.002189] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d4491-a46e-3ff1-cfe5-e475e89cc9d3" [ 991.002189] env[63197]: _type = "HttpNfcLease" [ 991.002189] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 991.002561] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 991.002561] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d4491-a46e-3ff1-cfe5-e475e89cc9d3" [ 991.002561] env[63197]: _type = "HttpNfcLease" [ 991.002561] env[63197]: }. {{(pid=63197) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 991.003382] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81b5cb8-082a-4a62-b51c-3a9bb9079f8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.011185] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204122d-4058-34df-579e-1e37901caf68/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 991.011368] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating HTTP connection to write to file with size = 31665664 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204122d-4058-34df-579e-1e37901caf68/disk-0.vmdk. {{(pid=63197) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 991.068363] env[63197]: DEBUG oslo_concurrency.lockutils [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.078191] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-121a92d7-838c-4e52-b08c-c27f414f4b37 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.080213] env[63197]: DEBUG oslo_concurrency.lockutils [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.080481] env[63197]: DEBUG nova.compute.manager [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.080667] env[63197]: DEBUG nova.compute.manager [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing instance network info cache due to event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.080883] env[63197]: DEBUG oslo_concurrency.lockutils [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.081038] env[63197]: DEBUG oslo_concurrency.lockutils [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.081205] env[63197]: DEBUG nova.network.neutron [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing network info cache for port 6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.099889] env[63197]: INFO nova.compute.manager [-] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Took 1.28 seconds to deallocate network for instance. [ 991.189614] env[63197]: INFO nova.compute.manager [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Took 13.43 seconds to build instance. [ 991.295019] env[63197]: DEBUG nova.compute.manager [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.295425] env[63197]: DEBUG nova.compute.manager [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing instance network info cache due to event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.295662] env[63197]: DEBUG oslo_concurrency.lockutils [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.295870] env[63197]: DEBUG oslo_concurrency.lockutils [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.296056] env[63197]: DEBUG nova.network.neutron [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.372824] env[63197]: DEBUG nova.scheduler.client.report [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.594296] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71546a47-1535-49a7-b07c-1d5d65dfd5e3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.615805] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.620375] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98beb861-c420-497e-a2f4-a41bca10bb73 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.632402] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance '64b944ea-2397-45eb-a0ed-9f57263db998' progress to 83 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.691047] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f3dadb1f-ae5d-463d-b446-c33f3dd54dbb tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.936s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.812773] env[63197]: DEBUG nova.network.neutron [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updated VIF entry in instance network info cache for port 6de5e286-38ff-452d-b567-9b6c956831d1. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.813246] env[63197]: DEBUG nova.network.neutron [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.879157] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.881584] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.266s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.881907] env[63197]: DEBUG nova.objects.instance [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lazy-loading 'resources' on Instance uuid c6686a09-3711-434b-b69f-a2518366fbed {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.908261] env[63197]: INFO nova.scheduler.client.report [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Deleted allocations for instance 4d12bac4-da42-42bd-9361-7015f3be9693 [ 992.099827] env[63197]: DEBUG nova.network.neutron [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updated VIF entry in instance network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.100268] env[63197]: DEBUG nova.network.neutron [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.139980] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 992.140316] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb9fbc47-9db9-4913-8690-0128cc6b3174 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.149285] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 992.149285] env[63197]: value = "task-1364475" [ 992.149285] env[63197]: _type = "Task" [ 992.149285] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.159122] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364475, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.231896] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Completed reading data from the image iterator. {{(pid=63197) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 992.232169] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204122d-4058-34df-579e-1e37901caf68/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 992.233413] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ae9656-8b5e-4db3-bc64-4cb6eb73e267 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.240590] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204122d-4058-34df-579e-1e37901caf68/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 992.240834] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204122d-4058-34df-579e-1e37901caf68/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 992.241124] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-03edc44a-a9af-4c3f-b9eb-534cb31929fe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.316607] env[63197]: DEBUG oslo_concurrency.lockutils [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.316891] env[63197]: DEBUG nova.compute.manager [req-7521095d-ee0f-43c1-a400-9d66af1e68cf req-e6bb8d32-1cb9-4f42-9f44-cbfd0431c34f service nova] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Received event network-vif-deleted-da5d991c-658a-47b4-b74f-ea5cb71fdbc6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.423748] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0a9aa46c-e20f-40ff-b520-af50618783bf tempest-ImagesOneServerTestJSON-1474694855 tempest-ImagesOneServerTestJSON-1474694855-project-member] Lock "4d12bac4-da42-42bd-9361-7015f3be9693" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.192s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.445461] env[63197]: DEBUG oslo_vmware.rw_handles [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5204122d-4058-34df-579e-1e37901caf68/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 992.445707] env[63197]: INFO nova.virt.vmwareapi.images [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Downloaded image file data 073c5ea5-cd55-4c9a-ab31-9918df041ed1 [ 992.446546] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb643c6-d05f-483d-af88-5ce8c857f920 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.469730] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21b64a92-8f82-41f6-8dd5-32d57b903475 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.500937] env[63197]: INFO nova.virt.vmwareapi.images [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] The imported VM was unregistered [ 992.503891] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Caching image {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 992.504154] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating directory with path [datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.506932] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19c2072a-a8ac-4fc2-adaa-2d4001bcfb65 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.521832] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created directory with path [datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 992.522063] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1/OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1.vmdk to [datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1/073c5ea5-cd55-4c9a-ab31-9918df041ed1.vmdk. {{(pid=63197) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 992.522336] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-0d5c7715-33c4-4061-b006-b270671ca4c3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.529380] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 992.529380] env[63197]: value = "task-1364477" [ 992.529380] env[63197]: _type = "Task" [ 992.529380] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.539010] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364477, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.544452] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f79a40-b70d-47c1-962e-8d1a3f04ddc6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.551294] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092b7010-79da-4da0-b124-f7e27e32bb5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.583353] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9213d0-60e9-4a21-907c-4adeef6d0891 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.597276] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704f1bcb-c226-4bca-828f-08ba1e0d00b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.603313] env[63197]: DEBUG oslo_concurrency.lockutils [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.603691] env[63197]: DEBUG nova.compute.manager [req-ab858f38-a715-4a1c-8023-b046a34fcb43 req-295f3950-a450-4da9-9c2d-06f99ccfec3b service nova] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Received event network-vif-deleted-0bdd3cb1-b50c-4791-9d45-3a1766a307da {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.612212] env[63197]: DEBUG nova.compute.provider_tree [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.615498] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.615971] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.616214] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.616413] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.616602] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.618654] env[63197]: INFO nova.compute.manager [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Terminating instance [ 992.620538] env[63197]: DEBUG nova.compute.manager [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 992.620737] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 992.621585] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54afa6fd-8d18-41ec-8b6e-d02c13e509a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.629952] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.630216] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03a0ce77-56f4-4447-8f6f-daf631aa5dfd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.636405] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 992.636405] env[63197]: value = "task-1364478" [ 992.636405] env[63197]: _type = "Task" [ 992.636405] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.644527] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.657177] env[63197]: DEBUG oslo_vmware.api [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364475, 'name': PowerOnVM_Task, 'duration_secs': 0.487198} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.657452] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.657649] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-320b4377-7a0f-4aa6-b832-8ce10a398327 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance '64b944ea-2397-45eb-a0ed-9f57263db998' progress to 100 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 993.041082] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364477, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.115576] env[63197]: DEBUG nova.scheduler.client.report [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.148909] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364478, 'name': PowerOffVM_Task, 'duration_secs': 0.233517} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.148909] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.148909] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 993.148909] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f576dd2f-1784-4b36-8cc3-92ff77722db9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.376288] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 993.376570] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 993.376811] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleting the datastore file [datastore2] 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.377262] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2eac3f89-38b7-4ff5-a55f-63a9710c0679 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.386039] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 993.386039] env[63197]: value = "task-1364480" [ 993.386039] env[63197]: _type = "Task" [ 993.386039] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.393917] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.540321] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364477, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.622610] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.739s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.656939] env[63197]: INFO nova.scheduler.client.report [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Deleted allocations for instance c6686a09-3711-434b-b69f-a2518366fbed [ 993.895521] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.041068] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364477, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.179121] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c5f43d4d-48af-445a-8cf1-bae23c4410ee tempest-SecurityGroupsTestJSON-129573762 tempest-SecurityGroupsTestJSON-129573762-project-member] Lock "c6686a09-3711-434b-b69f-a2518366fbed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.078s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.396015] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.542108] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364477, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.900714] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.043369] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364477, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.235768] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "64b944ea-2397-45eb-a0ed-9f57263db998" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.236177] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.236377] env[63197]: DEBUG nova.compute.manager [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Going to confirm migration 2 {{(pid=63197) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 995.398828] env[63197]: DEBUG oslo_vmware.api [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.87386} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.399113] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.399316] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.399531] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.399804] env[63197]: INFO nova.compute.manager [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Took 2.78 seconds to destroy the instance on the hypervisor. [ 995.400281] env[63197]: DEBUG oslo.service.loopingcall [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.400569] env[63197]: DEBUG nova.compute.manager [-] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.400681] env[63197]: DEBUG nova.network.neutron [-] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.543164] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364477, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.811872} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.543442] env[63197]: INFO nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1/OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1.vmdk to [datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1/073c5ea5-cd55-4c9a-ab31-9918df041ed1.vmdk. [ 995.543640] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Cleaning up location [datastore2] OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 995.543808] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_2ec35835-3e37-4321-87b5-f53870ae66b1 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.544090] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26bebc5c-76f3-40e2-8de6-9a3a01891b6d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.550522] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 995.550522] env[63197]: value = "task-1364481" [ 995.550522] env[63197]: _type = "Task" [ 995.550522] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.558202] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364481, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.783599] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.783887] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.784161] env[63197]: DEBUG nova.network.neutron [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.784424] env[63197]: DEBUG nova.objects.instance [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lazy-loading 'info_cache' on Instance uuid 64b944ea-2397-45eb-a0ed-9f57263db998 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.818270] env[63197]: DEBUG nova.compute.manager [req-5fe595cc-de43-4c9a-b519-610fbd10787c req-8c6c76d1-3144-43b1-a0c4-69e41f9133f9 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Received event network-vif-deleted-ed419ac7-cefa-4394-ad05-c3c85c6111e3 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.818580] env[63197]: INFO nova.compute.manager [req-5fe595cc-de43-4c9a-b519-610fbd10787c req-8c6c76d1-3144-43b1-a0c4-69e41f9133f9 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Neutron deleted interface ed419ac7-cefa-4394-ad05-c3c85c6111e3; detaching it from the instance and deleting it from the info cache [ 995.818723] env[63197]: DEBUG nova.network.neutron [req-5fe595cc-de43-4c9a-b519-610fbd10787c req-8c6c76d1-3144-43b1-a0c4-69e41f9133f9 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.063405] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364481, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24282} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.063762] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 996.063870] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1/073c5ea5-cd55-4c9a-ab31-9918df041ed1.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.064140] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1/073c5ea5-cd55-4c9a-ab31-9918df041ed1.vmdk to [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.064407] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c23e501-ea86-4202-8be4-1c6afc32e8df {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.071259] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 996.071259] env[63197]: value = "task-1364482" [ 996.071259] env[63197]: _type = "Task" [ 996.071259] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.079617] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.291080] env[63197]: DEBUG nova.network.neutron [-] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.324151] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0d83a64-29dd-459f-9196-c53d77c2b51f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.334365] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b61fe1-b68f-4ea9-b7b8-19b0cd501b4b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.362332] env[63197]: DEBUG nova.compute.manager [req-5fe595cc-de43-4c9a-b519-610fbd10787c req-8c6c76d1-3144-43b1-a0c4-69e41f9133f9 service nova] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Detach interface failed, port_id=ed419ac7-cefa-4394-ad05-c3c85c6111e3, reason: Instance 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 996.581229] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364482, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.793551] env[63197]: INFO nova.compute.manager [-] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Took 1.39 seconds to deallocate network for instance. [ 997.084480] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364482, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.199781] env[63197]: DEBUG nova.network.neutron [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance_info_cache with network_info: [{"id": "cbc85003-5324-451a-b9ba-25b9581079c7", "address": "fa:16:3e:21:a3:85", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcbc85003-53", "ovs_interfaceid": "cbc85003-5324-451a-b9ba-25b9581079c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.300295] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.300661] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.300906] env[63197]: DEBUG nova.objects.instance [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lazy-loading 'resources' on Instance uuid 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.585968] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364482, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.703348] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-64b944ea-2397-45eb-a0ed-9f57263db998" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.703767] env[63197]: DEBUG nova.objects.instance [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lazy-loading 'migration_context' on Instance uuid 64b944ea-2397-45eb-a0ed-9f57263db998 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.949162] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db82a687-129b-4db7-9946-276e276bc54e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.957214] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c640c4ab-48e2-4df5-a0ad-d242931cc5a0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.990509] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96121b5f-af3b-4aa3-a3c8-37c522b014d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.998754] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d929b0c-9cef-41b3-8af7-d55bcdaa968e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.013126] env[63197]: DEBUG nova.compute.provider_tree [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.084058] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364482, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.207203] env[63197]: DEBUG nova.objects.base [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Object Instance<64b944ea-2397-45eb-a0ed-9f57263db998> lazy-loaded attributes: info_cache,migration_context {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 998.208502] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e340a799-5ebd-4e34-af5d-25a5a26bf78d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.230216] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abf021af-be61-4441-ba35-36ebebe6e917 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.237434] env[63197]: DEBUG oslo_vmware.api [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 998.237434] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a355eb-4bda-272a-0e9d-45a24cf54867" [ 998.237434] env[63197]: _type = "Task" [ 998.237434] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.246923] env[63197]: DEBUG oslo_vmware.api [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a355eb-4bda-272a-0e9d-45a24cf54867, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.520027] env[63197]: DEBUG nova.scheduler.client.report [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.585497] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364482, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.749686] env[63197]: DEBUG oslo_vmware.api [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a355eb-4bda-272a-0e9d-45a24cf54867, 'name': SearchDatastore_Task, 'duration_secs': 0.01376} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.750095] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.930861] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 998.931186] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 999.023375] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.026349] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.276s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.047385] env[63197]: INFO nova.scheduler.client.report [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted allocations for instance 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa [ 999.085286] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364482, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.74718} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.085884] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/073c5ea5-cd55-4c9a-ab31-9918df041ed1/073c5ea5-cd55-4c9a-ab31-9918df041ed1.vmdk to [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.086264] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e719d6-1d14-4f73-986f-0bff8c76d985 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.108264] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk or device None with type streamOptimized {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.109423] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a568ab9-f84c-4ece-b196-c6e1a60febc5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.128445] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 999.128445] env[63197]: value = "task-1364483" [ 999.128445] env[63197]: _type = "Task" [ 999.128445] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.136170] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364483, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.438309] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 999.439672] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 999.439672] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Rebuilding the list of instances to heal {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 999.559684] env[63197]: DEBUG oslo_concurrency.lockutils [None req-08662940-53ed-4f04-b839-016d7b9093e9 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "430a73fa-ae22-45dd-a5c7-8c5ed616ebaa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.942s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.638659] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364483, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.641964] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017cb3d3-6a3c-4240-840a-512fdd40b02d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.649250] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c8b456-0541-4748-a979-ace4f8785233 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.681262] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0427653-7e51-439d-b362-ddbe345beb8f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.689157] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb64180-ec10-49ff-9c73-bfe72209728a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.704043] env[63197]: DEBUG nova.compute.provider_tree [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.977043] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.977170] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquired lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.977235] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Forcefully refreshing network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 999.977425] env[63197]: DEBUG nova.objects.instance [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lazy-loading 'info_cache' on Instance uuid 638ef9c9-253b-4958-a660-6c1801408a51 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.140666] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364483, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.149237] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.149893] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.208452] env[63197]: DEBUG nova.scheduler.client.report [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.641862] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364483, 'name': ReconfigVM_Task, 'duration_secs': 1.031468} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.642213] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3/180e0da7-f7ee-4fcd-be95-c2bf679278d3.vmdk or device None with type streamOptimized {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.643657] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'size': 0, 'boot_index': 0, 'disk_bus': None, 'encryption_secret_uuid': None, 'guest_format': None, 'encryption_options': None, 'device_name': '/dev/sda', 'device_type': 'disk', 'encryption_format': None, 'encrypted': False, 'image_id': 'f760fedc-0b5b-4c56-acbe-239b47c945e8'}], 'ephemerals': [], 'block_device_mapping': [{'boot_index': None, 'disk_bus': None, 'attachment_id': '62e695c2-8e3c-40a8-b76b-c0c8c6f6c20e', 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290384', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'name': 'volume-652f45ea-384f-474b-9fd4-23d569d7d473', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '180e0da7-f7ee-4fcd-be95-c2bf679278d3', 'attached_at': '', 'detached_at': '', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'serial': '652f45ea-384f-474b-9fd4-23d569d7d473'}, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63197) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1000.643875] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1000.644092] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290384', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'name': 'volume-652f45ea-384f-474b-9fd4-23d569d7d473', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '180e0da7-f7ee-4fcd-be95-c2bf679278d3', 'attached_at': '', 'detached_at': '', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'serial': '652f45ea-384f-474b-9fd4-23d569d7d473'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1000.644929] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023f953d-87b7-4517-afca-76d2e183d504 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.664604] env[63197]: DEBUG nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1000.670798] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45fac10-9146-4c14-8be5-a241a1b6375c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.709998] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] volume-652f45ea-384f-474b-9fd4-23d569d7d473/volume-652f45ea-384f-474b-9fd4-23d569d7d473.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.710372] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57c41ea8-87b7-4e23-b7a2-c8b9018de0ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.733150] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1000.733150] env[63197]: value = "task-1364484" [ 1000.733150] env[63197]: _type = "Task" [ 1000.733150] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.741680] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364484, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.916524] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.916621] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.196714] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.230348] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.205s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.233977] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.037s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.235432] env[63197]: INFO nova.compute.claims [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1001.246548] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364484, 'name': ReconfigVM_Task, 'duration_secs': 0.503416} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.246799] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfigured VM instance instance-00000049 to attach disk [datastore1] volume-652f45ea-384f-474b-9fd4-23d569d7d473/volume-652f45ea-384f-474b-9fd4-23d569d7d473.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.254018] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01c68531-c0ad-4caa-85b5-cb2bcddcebac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.276912] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1001.276912] env[63197]: value = "task-1364485" [ 1001.276912] env[63197]: _type = "Task" [ 1001.276912] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.288345] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364485, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.314384] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "d8919322-1fa7-4fc7-9380-a8b7d283b050" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.314960] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.419415] env[63197]: DEBUG nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1001.506769] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.507021] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.514354] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-4672f595-e1f6-4400-b5a1-065598584980-44d19552-d824-4f52-93fa-832e41750693" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.514705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-4672f595-e1f6-4400-b5a1-065598584980-44d19552-d824-4f52-93fa-832e41750693" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.515145] env[63197]: DEBUG nova.objects.instance [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'flavor' on Instance uuid 4672f595-e1f6-4400-b5a1-065598584980 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.756830] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updating instance_info_cache with network_info: [{"id": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "address": "fa:16:3e:7e:b1:97", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.152", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape927a6cd-69", "ovs_interfaceid": "e927a6cd-692a-4cce-b2f6-8d321333e34f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.788264] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364485, 'name': ReconfigVM_Task, 'duration_secs': 0.364372} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.788531] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290384', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'name': 'volume-652f45ea-384f-474b-9fd4-23d569d7d473', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '180e0da7-f7ee-4fcd-be95-c2bf679278d3', 'attached_at': '', 'detached_at': '', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'serial': '652f45ea-384f-474b-9fd4-23d569d7d473'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1001.789189] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b1ec1c4-df72-4c39-88fd-0b56f7ea91cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.796297] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1001.796297] env[63197]: value = "task-1364486" [ 1001.796297] env[63197]: _type = "Task" [ 1001.796297] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.802338] env[63197]: INFO nova.scheduler.client.report [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted allocation for migration 5d03443b-9ea8-4897-a373-50aca57af1de [ 1001.807246] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364486, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.817689] env[63197]: DEBUG nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1001.940685] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.008968] env[63197]: DEBUG nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1002.172313] env[63197]: DEBUG nova.objects.instance [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'pci_requests' on Instance uuid 4672f595-e1f6-4400-b5a1-065598584980 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.260053] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Releasing lock "refresh_cache-638ef9c9-253b-4958-a660-6c1801408a51" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.260193] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updated the network info_cache for instance {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1002.260381] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1002.260669] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1002.260771] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1002.260930] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1002.261125] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1002.261797] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1002.261797] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1002.261797] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1002.307019] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364486, 'name': Rename_Task, 'duration_secs': 0.156596} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.309679] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1002.310607] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b3db401-df1b-4c44-bd0a-705cb9b70586 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.312956] env[63197]: DEBUG oslo_concurrency.lockutils [None req-366b30b1-a001-4006-8246-ef5489de1947 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.077s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.319584] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1002.319584] env[63197]: value = "task-1364487" [ 1002.319584] env[63197]: _type = "Task" [ 1002.319584] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.335147] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364487, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.344093] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.418130] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3208b237-21ca-4481-b7ce-d4bdc6d5a6ea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.425323] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22aa2a8-8578-4a44-b2e5-7e78e520b467 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.456263] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6a0824-81c7-4cc6-a7b0-dfbbf9392249 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.463775] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c080fe0-723f-465b-b645-4bdb9d8fe097 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.476964] env[63197]: DEBUG nova.compute.provider_tree [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.534440] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.677900] env[63197]: DEBUG nova.objects.base [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Object Instance<4672f595-e1f6-4400-b5a1-065598584980> lazy-loaded attributes: flavor,pci_requests {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1002.678174] env[63197]: DEBUG nova.network.neutron [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1002.755842] env[63197]: DEBUG nova.policy [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1002.757896] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "64b944ea-2397-45eb-a0ed-9f57263db998" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.758182] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.758389] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.758596] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.758773] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.761051] env[63197]: INFO nova.compute.manager [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Terminating instance [ 1002.762506] env[63197]: DEBUG nova.compute.manager [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1002.762727] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.763798] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8776b46-021a-44ca-b12a-8c568eb7cee3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.766832] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.771705] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.771946] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e816c60-a0be-4dc9-b60c-d1f7623eb6da {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.778494] env[63197]: DEBUG oslo_vmware.api [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1002.778494] env[63197]: value = "task-1364488" [ 1002.778494] env[63197]: _type = "Task" [ 1002.778494] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.785900] env[63197]: DEBUG oslo_vmware.api [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364488, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.830043] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364487, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.980136] env[63197]: DEBUG nova.scheduler.client.report [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.225435] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "7f9ad29e-9cb9-4575-8dce-081c02767af5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.225635] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.288375] env[63197]: DEBUG oslo_vmware.api [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364488, 'name': PowerOffVM_Task, 'duration_secs': 0.225593} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.288652] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.288878] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.289158] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-080ad9ba-4dd2-469f-9ac5-ace281e31787 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.329786] env[63197]: DEBUG oslo_vmware.api [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364487, 'name': PowerOnVM_Task, 'duration_secs': 0.729142} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.330067] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.409884] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.410143] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.410332] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleting the datastore file [datastore1] 64b944ea-2397-45eb-a0ed-9f57263db998 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.410709] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a631ad6c-5f74-405f-8535-0ac315e66d95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.416903] env[63197]: DEBUG oslo_vmware.api [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1003.416903] env[63197]: value = "task-1364490" [ 1003.416903] env[63197]: _type = "Task" [ 1003.416903] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.424833] env[63197]: DEBUG oslo_vmware.api [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.429161] env[63197]: DEBUG nova.compute.manager [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1003.429986] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897484b1-4b60-48c9-a5a4-cf21d061c6b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.485245] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.251s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.485725] env[63197]: DEBUG nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1003.488422] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.548s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.490336] env[63197]: INFO nova.compute.claims [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1003.657140] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "8e5ef1b0-7532-498a-84c2-189274a36c50" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.657462] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.657716] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "8e5ef1b0-7532-498a-84c2-189274a36c50-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.657953] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.658178] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.660222] env[63197]: INFO nova.compute.manager [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Terminating instance [ 1003.662083] env[63197]: DEBUG nova.compute.manager [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.662364] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.663336] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef02857-1e76-436f-af79-ca07bf88f722 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.671223] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.671449] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a75919c8-766e-42c3-8386-1d796801e58a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.677181] env[63197]: DEBUG oslo_vmware.api [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1003.677181] env[63197]: value = "task-1364491" [ 1003.677181] env[63197]: _type = "Task" [ 1003.677181] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.684820] env[63197]: DEBUG oslo_vmware.api [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.729673] env[63197]: DEBUG nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1003.926507] env[63197]: DEBUG oslo_vmware.api [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364490, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267927} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.926866] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.926971] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.927181] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.927358] env[63197]: INFO nova.compute.manager [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1003.927603] env[63197]: DEBUG oslo.service.loopingcall [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.927803] env[63197]: DEBUG nova.compute.manager [-] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.927895] env[63197]: DEBUG nova.network.neutron [-] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1003.945029] env[63197]: DEBUG oslo_concurrency.lockutils [None req-efb84b75-3005-448a-bb8d-33d6bece30e9 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.684s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.994824] env[63197]: DEBUG nova.compute.utils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.999025] env[63197]: DEBUG nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1003.999088] env[63197]: DEBUG nova.network.neutron [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1004.071992] env[63197]: DEBUG nova.policy [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a8e87b64fe46d7ab7e570d57611119', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ceb1016d6d34bff8880dca42d495377', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1004.187972] env[63197]: DEBUG oslo_vmware.api [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364491, 'name': PowerOffVM_Task, 'duration_secs': 0.205713} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.188456] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.188838] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1004.189257] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d75d495-63b2-41eb-8f7c-c373ca477f95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.254312] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.254312] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.254312] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleting the datastore file [datastore2] 8e5ef1b0-7532-498a-84c2-189274a36c50 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.254312] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb892c3c-abd3-4bcd-a642-230a4e5d4e18 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.262806] env[63197]: DEBUG oslo_vmware.api [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1004.262806] env[63197]: value = "task-1364493" [ 1004.262806] env[63197]: _type = "Task" [ 1004.262806] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.264167] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.277226] env[63197]: DEBUG oslo_vmware.api [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.284790] env[63197]: DEBUG nova.compute.manager [req-71d0962a-5791-4f41-bdc3-3469a6eb6de2 req-77af4868-eff5-4bea-84e9-6686c60113cd service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-vif-plugged-44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.284884] env[63197]: DEBUG oslo_concurrency.lockutils [req-71d0962a-5791-4f41-bdc3-3469a6eb6de2 req-77af4868-eff5-4bea-84e9-6686c60113cd service nova] Acquiring lock "4672f595-e1f6-4400-b5a1-065598584980-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.285073] env[63197]: DEBUG oslo_concurrency.lockutils [req-71d0962a-5791-4f41-bdc3-3469a6eb6de2 req-77af4868-eff5-4bea-84e9-6686c60113cd service nova] Lock "4672f595-e1f6-4400-b5a1-065598584980-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.286167] env[63197]: DEBUG oslo_concurrency.lockutils [req-71d0962a-5791-4f41-bdc3-3469a6eb6de2 req-77af4868-eff5-4bea-84e9-6686c60113cd service nova] Lock "4672f595-e1f6-4400-b5a1-065598584980-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.286167] env[63197]: DEBUG nova.compute.manager [req-71d0962a-5791-4f41-bdc3-3469a6eb6de2 req-77af4868-eff5-4bea-84e9-6686c60113cd service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] No waiting events found dispatching network-vif-plugged-44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.286167] env[63197]: WARNING nova.compute.manager [req-71d0962a-5791-4f41-bdc3-3469a6eb6de2 req-77af4868-eff5-4bea-84e9-6686c60113cd service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received unexpected event network-vif-plugged-44d19552-d824-4f52-93fa-832e41750693 for instance with vm_state active and task_state None. [ 1004.418544] env[63197]: DEBUG nova.network.neutron [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Successfully created port: 0284be0f-c3dd-4db2-8a6b-10ec63920553 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1004.497921] env[63197]: DEBUG nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1004.613109] env[63197]: DEBUG nova.network.neutron [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Successfully updated port: 44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1004.638135] env[63197]: DEBUG nova.compute.manager [req-b6724fea-5187-4301-9569-b784b0a31250 req-d6f02ed4-6fd3-45a5-a1a8-16b4eabed822 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Received event network-vif-deleted-cbc85003-5324-451a-b9ba-25b9581079c7 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.638350] env[63197]: INFO nova.compute.manager [req-b6724fea-5187-4301-9569-b784b0a31250 req-d6f02ed4-6fd3-45a5-a1a8-16b4eabed822 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Neutron deleted interface cbc85003-5324-451a-b9ba-25b9581079c7; detaching it from the instance and deleting it from the info cache [ 1004.638527] env[63197]: DEBUG nova.network.neutron [req-b6724fea-5187-4301-9569-b784b0a31250 req-d6f02ed4-6fd3-45a5-a1a8-16b4eabed822 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.670731] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f496aa-4d7f-4b8c-aa1a-478548261572 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.678174] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86dc61d-e288-42fe-bb63-9d11e9c231ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.706669] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06cee8f1-a41c-425c-8cca-3acdfa0b9750 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.713660] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8e92413-e86a-4513-b2d1-b37294d7236c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.726246] env[63197]: DEBUG nova.compute.provider_tree [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.776717] env[63197]: DEBUG oslo_vmware.api [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166453} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.777375] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.777375] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.777512] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.777599] env[63197]: INFO nova.compute.manager [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1004.778033] env[63197]: DEBUG oslo.service.loopingcall [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.778273] env[63197]: DEBUG nova.compute.manager [-] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.778451] env[63197]: DEBUG nova.network.neutron [-] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.805865] env[63197]: DEBUG nova.network.neutron [-] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.115633] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.115936] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.116028] env[63197]: DEBUG nova.network.neutron [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1005.143352] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-727876cb-28df-48c6-bed5-a772806396a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.154303] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a476af5-8413-4c14-8018-bee2cfdea36a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.181764] env[63197]: DEBUG nova.compute.manager [req-b6724fea-5187-4301-9569-b784b0a31250 req-d6f02ed4-6fd3-45a5-a1a8-16b4eabed822 service nova] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Detach interface failed, port_id=cbc85003-5324-451a-b9ba-25b9581079c7, reason: Instance 64b944ea-2397-45eb-a0ed-9f57263db998 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1005.229928] env[63197]: DEBUG nova.scheduler.client.report [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.309220] env[63197]: INFO nova.compute.manager [-] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Took 1.38 seconds to deallocate network for instance. [ 1005.511176] env[63197]: DEBUG nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1005.537855] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1005.538148] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1005.538317] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.538499] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1005.538647] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.538795] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1005.538999] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1005.539443] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1005.539653] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1005.539840] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1005.540043] env[63197]: DEBUG nova.virt.hardware [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1005.540954] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930537d8-247d-4084-9018-e940ca1384db {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.548929] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51165bcc-484e-4616-b83e-9797f43a21d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.649364] env[63197]: WARNING nova.network.neutron [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] 7ac08e03-1288-4a87-9a9c-257ba778c6d6 already exists in list: networks containing: ['7ac08e03-1288-4a87-9a9c-257ba778c6d6']. ignoring it [ 1005.738813] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.739442] env[63197]: DEBUG nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.745436] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.401s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.746644] env[63197]: INFO nova.compute.claims [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.777033] env[63197]: DEBUG nova.network.neutron [-] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.816548] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.961182] env[63197]: DEBUG nova.network.neutron [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Successfully updated port: 0284be0f-c3dd-4db2-8a6b-10ec63920553 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1006.017650] env[63197]: DEBUG nova.network.neutron [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "44d19552-d824-4f52-93fa-832e41750693", "address": "fa:16:3e:35:a8:2e", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d19552-d8", "ovs_interfaceid": "44d19552-d824-4f52-93fa-832e41750693", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.251057] env[63197]: DEBUG nova.compute.utils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1006.254305] env[63197]: DEBUG nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1006.254305] env[63197]: DEBUG nova.network.neutron [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1006.281024] env[63197]: INFO nova.compute.manager [-] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Took 1.50 seconds to deallocate network for instance. [ 1006.296584] env[63197]: DEBUG nova.policy [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12d00044b65c48f484ac9fea595d6179', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80138785cfdb4a2188ffb972da437646', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.415421] env[63197]: DEBUG nova.compute.manager [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-changed-44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.415658] env[63197]: DEBUG nova.compute.manager [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing instance network info cache due to event network-changed-44d19552-d824-4f52-93fa-832e41750693. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.415799] env[63197]: DEBUG oslo_concurrency.lockutils [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.465313] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-f39f3e80-0d96-4024-aec8-16e20e9bf13f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.465442] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-f39f3e80-0d96-4024-aec8-16e20e9bf13f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.466042] env[63197]: DEBUG nova.network.neutron [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.520675] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.521359] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.521542] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.521981] env[63197]: DEBUG oslo_concurrency.lockutils [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.522186] env[63197]: DEBUG nova.network.neutron [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing network info cache for port 44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.524987] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00753e02-0ac1-4f5a-b5d3-9b3c330d4271 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.543626] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.543847] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.544012] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.544248] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.544409] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.544561] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.544764] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.544965] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.545193] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.545283] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.549019] env[63197]: DEBUG nova.virt.hardware [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.552466] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Reconfiguring VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1006.555496] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-983f5e00-ede3-4bfb-b77c-ee2993efa095 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.577998] env[63197]: DEBUG oslo_vmware.api [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1006.577998] env[63197]: value = "task-1364494" [ 1006.577998] env[63197]: _type = "Task" [ 1006.577998] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.578762] env[63197]: DEBUG nova.network.neutron [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Successfully created port: 7a9f024c-f13b-437a-8b6c-ea3c7752044a {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.588932] env[63197]: DEBUG oslo_vmware.api [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364494, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.663533] env[63197]: DEBUG nova.compute.manager [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Received event network-vif-plugged-0284be0f-c3dd-4db2-8a6b-10ec63920553 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.663620] env[63197]: DEBUG oslo_concurrency.lockutils [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] Acquiring lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.663835] env[63197]: DEBUG oslo_concurrency.lockutils [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.664177] env[63197]: DEBUG oslo_concurrency.lockutils [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.664448] env[63197]: DEBUG nova.compute.manager [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] No waiting events found dispatching network-vif-plugged-0284be0f-c3dd-4db2-8a6b-10ec63920553 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1006.664637] env[63197]: WARNING nova.compute.manager [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Received unexpected event network-vif-plugged-0284be0f-c3dd-4db2-8a6b-10ec63920553 for instance with vm_state building and task_state spawning. [ 1006.664825] env[63197]: DEBUG nova.compute.manager [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Received event network-changed-0284be0f-c3dd-4db2-8a6b-10ec63920553 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.665077] env[63197]: DEBUG nova.compute.manager [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Refreshing instance network info cache due to event network-changed-0284be0f-c3dd-4db2-8a6b-10ec63920553. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.665273] env[63197]: DEBUG oslo_concurrency.lockutils [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] Acquiring lock "refresh_cache-f39f3e80-0d96-4024-aec8-16e20e9bf13f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.754996] env[63197]: DEBUG nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.788244] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.925907] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1db93b-a854-4c5a-84c1-e689471a77a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.936061] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74628e13-05b5-4a19-a2ce-49b27ad84ec6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.972776] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f6ed64-947e-41ea-8d8d-4472a1ce19ce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.981934] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de32504-e75d-4564-86b1-f053352d02f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.995753] env[63197]: DEBUG nova.compute.provider_tree [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.002236] env[63197]: DEBUG nova.network.neutron [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1007.090302] env[63197]: DEBUG oslo_vmware.api [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364494, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.239172] env[63197]: DEBUG nova.network.neutron [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Updating instance_info_cache with network_info: [{"id": "0284be0f-c3dd-4db2-8a6b-10ec63920553", "address": "fa:16:3e:83:3a:c5", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0284be0f-c3", "ovs_interfaceid": "0284be0f-c3dd-4db2-8a6b-10ec63920553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.293800] env[63197]: DEBUG nova.network.neutron [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updated VIF entry in instance network info cache for port 44d19552-d824-4f52-93fa-832e41750693. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.294264] env[63197]: DEBUG nova.network.neutron [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "44d19552-d824-4f52-93fa-832e41750693", "address": "fa:16:3e:35:a8:2e", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d19552-d8", "ovs_interfaceid": "44d19552-d824-4f52-93fa-832e41750693", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.500156] env[63197]: DEBUG nova.scheduler.client.report [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.591565] env[63197]: DEBUG oslo_vmware.api [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364494, 'name': ReconfigVM_Task, 'duration_secs': 0.575039} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.592126] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.592349] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Reconfigured VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1007.742218] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-f39f3e80-0d96-4024-aec8-16e20e9bf13f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.742494] env[63197]: DEBUG nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Instance network_info: |[{"id": "0284be0f-c3dd-4db2-8a6b-10ec63920553", "address": "fa:16:3e:83:3a:c5", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0284be0f-c3", "ovs_interfaceid": "0284be0f-c3dd-4db2-8a6b-10ec63920553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1007.742836] env[63197]: DEBUG oslo_concurrency.lockutils [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] Acquired lock "refresh_cache-f39f3e80-0d96-4024-aec8-16e20e9bf13f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.743072] env[63197]: DEBUG nova.network.neutron [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Refreshing network info cache for port 0284be0f-c3dd-4db2-8a6b-10ec63920553 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.744757] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:3a:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0284be0f-c3dd-4db2-8a6b-10ec63920553', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.752812] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating folder: Project (0ceb1016d6d34bff8880dca42d495377). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1007.755886] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66c8944c-8992-44cb-9df9-1f4672ee9fc1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.766834] env[63197]: DEBUG nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.769550] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created folder: Project (0ceb1016d6d34bff8880dca42d495377) in parent group-v290286. [ 1007.769887] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating folder: Instances. Parent ref: group-v290389. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1007.769987] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9d8319b-f7b3-411e-87f9-7023a091b6bc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.778658] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created folder: Instances in parent group-v290389. [ 1007.778901] env[63197]: DEBUG oslo.service.loopingcall [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.779119] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1007.779315] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d4eed3c-fb06-4eee-8007-30abc9c3cb1a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.798300] env[63197]: DEBUG oslo_concurrency.lockutils [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.799153] env[63197]: DEBUG nova.compute.manager [req-5e9e4a3b-b0dd-4f0d-8323-9940b752344e req-a06a34fb-2222-4a43-b46c-d7dc6594a0be service nova] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Received event network-vif-deleted-fa39c111-5e72-43e8-b764-d2072c0c3d40 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.804283] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.804283] env[63197]: value = "task-1364497" [ 1007.804283] env[63197]: _type = "Task" [ 1007.804283] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.806349] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.806584] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.806832] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.807185] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.807869] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.807869] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.808104] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.808375] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.808672] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.808964] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.809272] env[63197]: DEBUG nova.virt.hardware [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.810540] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b8a62e-d8c4-45c5-87ad-fb4a4b2fa550 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.832324] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364497, 'name': CreateVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.835036] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a74d3f-314e-40bf-b274-6ea47e9f54e6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.982073] env[63197]: DEBUG nova.network.neutron [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Updated VIF entry in instance network info cache for port 0284be0f-c3dd-4db2-8a6b-10ec63920553. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.982461] env[63197]: DEBUG nova.network.neutron [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Updating instance_info_cache with network_info: [{"id": "0284be0f-c3dd-4db2-8a6b-10ec63920553", "address": "fa:16:3e:83:3a:c5", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0284be0f-c3", "ovs_interfaceid": "0284be0f-c3dd-4db2-8a6b-10ec63920553", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.008785] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.009473] env[63197]: DEBUG nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1008.013335] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.479s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.018020] env[63197]: INFO nova.compute.claims [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.096064] env[63197]: DEBUG nova.network.neutron [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Successfully updated port: 7a9f024c-f13b-437a-8b6c-ea3c7752044a {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.099028] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f54f733b-7878-4c69-8e27-25509f43e016 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-4672f595-e1f6-4400-b5a1-065598584980-44d19552-d824-4f52-93fa-832e41750693" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.584s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.319669] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364497, 'name': CreateVM_Task, 'duration_secs': 0.362053} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.320046] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1008.320466] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.320651] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.320969] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1008.321236] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-813ee625-f6fc-4667-b4e9-fa84eb7707ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.325556] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1008.325556] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aa45c1-530d-a7cf-081f-1086d2381ed9" [ 1008.325556] env[63197]: _type = "Task" [ 1008.325556] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.333055] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aa45c1-530d-a7cf-081f-1086d2381ed9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.438956] env[63197]: DEBUG nova.compute.manager [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Received event network-vif-plugged-7a9f024c-f13b-437a-8b6c-ea3c7752044a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.439182] env[63197]: DEBUG oslo_concurrency.lockutils [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] Acquiring lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.439448] env[63197]: DEBUG oslo_concurrency.lockutils [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.439562] env[63197]: DEBUG oslo_concurrency.lockutils [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.439735] env[63197]: DEBUG nova.compute.manager [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] No waiting events found dispatching network-vif-plugged-7a9f024c-f13b-437a-8b6c-ea3c7752044a {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1008.439903] env[63197]: WARNING nova.compute.manager [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Received unexpected event network-vif-plugged-7a9f024c-f13b-437a-8b6c-ea3c7752044a for instance with vm_state building and task_state spawning. [ 1008.440204] env[63197]: DEBUG nova.compute.manager [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Received event network-changed-7a9f024c-f13b-437a-8b6c-ea3c7752044a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.440396] env[63197]: DEBUG nova.compute.manager [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Refreshing instance network info cache due to event network-changed-7a9f024c-f13b-437a-8b6c-ea3c7752044a. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.440586] env[63197]: DEBUG oslo_concurrency.lockutils [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] Acquiring lock "refresh_cache-cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.440733] env[63197]: DEBUG oslo_concurrency.lockutils [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] Acquired lock "refresh_cache-cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.440895] env[63197]: DEBUG nova.network.neutron [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Refreshing network info cache for port 7a9f024c-f13b-437a-8b6c-ea3c7752044a {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.484853] env[63197]: DEBUG oslo_concurrency.lockutils [req-b136b7e1-93aa-4984-91d0-525d4133bb83 req-e4952650-1ffe-4b1d-a98b-08b55f933b4c service nova] Releasing lock "refresh_cache-f39f3e80-0d96-4024-aec8-16e20e9bf13f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.519309] env[63197]: DEBUG nova.compute.utils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1008.522878] env[63197]: DEBUG nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1008.523052] env[63197]: DEBUG nova.network.neutron [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1008.602831] env[63197]: DEBUG nova.policy [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91e4c3c1b1c44ae5be850e278df04408', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ece2f02eca2440986f0f5c720fb7d1e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1008.605085] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "refresh_cache-cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.836742] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aa45c1-530d-a7cf-081f-1086d2381ed9, 'name': SearchDatastore_Task, 'duration_secs': 0.00979} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.837086] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.837320] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.837548] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.837694] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.837866] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.838336] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-162f8f77-c611-40de-af11-fc2eac44607f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.847225] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.847419] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1008.848216] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8523060e-1db0-4bd6-bca9-7bcfd1b4c5ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.854813] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1008.854813] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d4e0e9-db69-a76c-e68a-3b2210e82843" [ 1008.854813] env[63197]: _type = "Task" [ 1008.854813] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.862812] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d4e0e9-db69-a76c-e68a-3b2210e82843, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.932622] env[63197]: DEBUG nova.network.neutron [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Successfully created port: ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.973011] env[63197]: DEBUG nova.network.neutron [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1009.026578] env[63197]: DEBUG nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1009.047437] env[63197]: DEBUG nova.network.neutron [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.199562] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4876dfeb-3b0e-4793-819b-5598163d7a89 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.207257] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f393110-df16-4a2a-b0d4-f7c5fa62a31a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.250405] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad27dd2-020f-4e01-8522-a79896eec000 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.258429] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c211375-dcde-4ae7-b279-b7da85869095 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.271775] env[63197]: DEBUG nova.compute.provider_tree [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.368025] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52d4e0e9-db69-a76c-e68a-3b2210e82843, 'name': SearchDatastore_Task, 'duration_secs': 0.009179} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.368025] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff03304d-8495-4483-b16c-108880cfd01e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.371458] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1009.371458] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524725dd-dd3f-5ded-7d79-73d7ade09f20" [ 1009.371458] env[63197]: _type = "Task" [ 1009.371458] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.379596] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524725dd-dd3f-5ded-7d79-73d7ade09f20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.551458] env[63197]: DEBUG oslo_concurrency.lockutils [req-772b4a28-592c-4b4e-ae4b-7f158e222fd7 req-2c570f8c-c03e-4270-bcbd-bb83bedf6e39 service nova] Releasing lock "refresh_cache-cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.551458] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "refresh_cache-cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.551458] env[63197]: DEBUG nova.network.neutron [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1009.776065] env[63197]: DEBUG nova.scheduler.client.report [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.810879] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-4672f595-e1f6-4400-b5a1-065598584980-44d19552-d824-4f52-93fa-832e41750693" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.811326] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-4672f595-e1f6-4400-b5a1-065598584980-44d19552-d824-4f52-93fa-832e41750693" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.883232] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]524725dd-dd3f-5ded-7d79-73d7ade09f20, 'name': SearchDatastore_Task, 'duration_secs': 0.009135} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.883459] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.883752] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] f39f3e80-0d96-4024-aec8-16e20e9bf13f/f39f3e80-0d96-4024-aec8-16e20e9bf13f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1009.883969] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b159248-5510-4230-a949-42bd156c89c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.890802] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1009.890802] env[63197]: value = "task-1364498" [ 1009.890802] env[63197]: _type = "Task" [ 1009.890802] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.898199] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.038381] env[63197]: DEBUG nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1010.063683] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.063933] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.064112] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.064339] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.064548] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.064722] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.064940] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.065129] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.065365] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.065602] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.065860] env[63197]: DEBUG nova.virt.hardware [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.066783] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58ff48f-b77d-43dd-b3ff-688261fbb050 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.074868] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7238c1bf-a335-41a0-95de-0ffcc8944184 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.080806] env[63197]: DEBUG nova.network.neutron [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1010.229145] env[63197]: DEBUG nova.network.neutron [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Updating instance_info_cache with network_info: [{"id": "7a9f024c-f13b-437a-8b6c-ea3c7752044a", "address": "fa:16:3e:0e:0b:bc", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9f024c-f1", "ovs_interfaceid": "7a9f024c-f13b-437a-8b6c-ea3c7752044a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.280934] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.281573] env[63197]: DEBUG nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1010.284761] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.518s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.285062] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.285311] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1010.285737] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.022s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.287465] env[63197]: INFO nova.compute.claims [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.291965] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76abeb8-ae8b-44d0-a822-1762969035c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.303025] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50789800-9195-4660-a0dd-a2d820fd6cc8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.319353] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.319585] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.320802] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7102ed4b-7f51-47cd-9d9d-9ac89ae4ca63 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.324186] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385a0a96-0295-4750-9c5e-f2f86f6a9e91 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.345559] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47439990-401d-484c-a72a-ee1c25d790cd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.350253] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e434f7be-f86c-4778-b75d-ff21e6d25059 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.382893] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Reconfiguring VM to detach interface {{(pid=63197) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1010.408739] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd4cb387-42dc-450c-ab40-875a847c3eb5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.423080] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180285MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1010.423256] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.435213] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364498, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.436880] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1010.436880] env[63197]: value = "task-1364499" [ 1010.436880] env[63197]: _type = "Task" [ 1010.436880] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.448617] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.470937] env[63197]: DEBUG nova.compute.manager [req-72542cfc-d990-4dcd-8b7e-571ff085fe83 req-11cfe6fb-3a27-43ed-a74e-7d1e9d17d0ac service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Received event network-vif-plugged-ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1010.471235] env[63197]: DEBUG oslo_concurrency.lockutils [req-72542cfc-d990-4dcd-8b7e-571ff085fe83 req-11cfe6fb-3a27-43ed-a74e-7d1e9d17d0ac service nova] Acquiring lock "d8919322-1fa7-4fc7-9380-a8b7d283b050-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.471497] env[63197]: DEBUG oslo_concurrency.lockutils [req-72542cfc-d990-4dcd-8b7e-571ff085fe83 req-11cfe6fb-3a27-43ed-a74e-7d1e9d17d0ac service nova] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.471716] env[63197]: DEBUG oslo_concurrency.lockutils [req-72542cfc-d990-4dcd-8b7e-571ff085fe83 req-11cfe6fb-3a27-43ed-a74e-7d1e9d17d0ac service nova] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.471897] env[63197]: DEBUG nova.compute.manager [req-72542cfc-d990-4dcd-8b7e-571ff085fe83 req-11cfe6fb-3a27-43ed-a74e-7d1e9d17d0ac service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] No waiting events found dispatching network-vif-plugged-ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.472120] env[63197]: WARNING nova.compute.manager [req-72542cfc-d990-4dcd-8b7e-571ff085fe83 req-11cfe6fb-3a27-43ed-a74e-7d1e9d17d0ac service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Received unexpected event network-vif-plugged-ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 for instance with vm_state building and task_state spawning. [ 1010.498825] env[63197]: DEBUG nova.network.neutron [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Successfully updated port: ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1010.731672] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "refresh_cache-cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.731988] env[63197]: DEBUG nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Instance network_info: |[{"id": "7a9f024c-f13b-437a-8b6c-ea3c7752044a", "address": "fa:16:3e:0e:0b:bc", "network": {"id": "c3a8cf44-c126-4b64-80f9-25ee157d430a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1127345199-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "80138785cfdb4a2188ffb972da437646", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d062877-0b23-4965-908b-f585f25f3bf1", "external-id": "nsx-vlan-transportzone-523", "segmentation_id": 523, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9f024c-f1", "ovs_interfaceid": "7a9f024c-f13b-437a-8b6c-ea3c7752044a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1010.732443] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:0b:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d062877-0b23-4965-908b-f585f25f3bf1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a9f024c-f13b-437a-8b6c-ea3c7752044a', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.740760] env[63197]: DEBUG oslo.service.loopingcall [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.741031] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1010.741318] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d504e95-9e03-4028-9d41-48e800a84598 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.761530] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1010.761530] env[63197]: value = "task-1364500" [ 1010.761530] env[63197]: _type = "Task" [ 1010.761530] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.769366] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364500, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.794054] env[63197]: DEBUG nova.compute.utils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.794785] env[63197]: DEBUG nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.794948] env[63197]: DEBUG nova.network.neutron [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1010.852657] env[63197]: DEBUG nova.policy [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91e4c3c1b1c44ae5be850e278df04408', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ece2f02eca2440986f0f5c720fb7d1e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.932602] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.642972} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.932884] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] f39f3e80-0d96-4024-aec8-16e20e9bf13f/f39f3e80-0d96-4024-aec8-16e20e9bf13f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1010.933119] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1010.933376] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3bceb4b-0232-41c5-9338-be1368583a8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.942212] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1010.942212] env[63197]: value = "task-1364501" [ 1010.942212] env[63197]: _type = "Task" [ 1010.942212] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.948468] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.953141] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364501, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.002058] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "refresh_cache-d8919322-1fa7-4fc7-9380-a8b7d283b050" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.002232] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired lock "refresh_cache-d8919322-1fa7-4fc7-9380-a8b7d283b050" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.002387] env[63197]: DEBUG nova.network.neutron [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1011.116227] env[63197]: DEBUG nova.network.neutron [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Successfully created port: 9161195e-f24b-453c-b121-a6a965441223 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1011.271641] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364500, 'name': CreateVM_Task, 'duration_secs': 0.345805} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.271812] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.272501] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.272686] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.273491] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.273804] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4795b545-7f1c-4a50-9554-9b33cdc2a4a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.278351] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1011.278351] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eb422e-4539-20e4-2897-fd127a683316" [ 1011.278351] env[63197]: _type = "Task" [ 1011.278351] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.286511] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eb422e-4539-20e4-2897-fd127a683316, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.298108] env[63197]: DEBUG nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1011.442312] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed25d7d4-7e49-4a41-9be0-acaa6bb342b6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.450890] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.453770] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b7fc98-17fa-405e-9a19-d5a694f5069e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.459164] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364501, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071794} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.459683] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.460394] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c385cb9b-0b0c-4e3e-b635-3f07890c51d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.489116] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099c63ac-5bf2-4ab3-a4b9-092cb6b13df0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.509083] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] f39f3e80-0d96-4024-aec8-16e20e9bf13f/f39f3e80-0d96-4024-aec8-16e20e9bf13f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.511301] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9df4c6a6-fe3c-467f-8dc7-1a5ccd715e7e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.528474] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cc4dbb-0bd5-42c8-8f10-f692d0240db9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.533198] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1011.533198] env[63197]: value = "task-1364502" [ 1011.533198] env[63197]: _type = "Task" [ 1011.533198] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.543453] env[63197]: DEBUG nova.compute.provider_tree [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.548991] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364502, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.553244] env[63197]: DEBUG nova.network.neutron [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1011.676507] env[63197]: DEBUG nova.network.neutron [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Updating instance_info_cache with network_info: [{"id": "ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52", "address": "fa:16:3e:ee:64:45", "network": {"id": "72ef7714-d486-4e2b-a3e7-c52ede2f7dd6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.56", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ba0c482e322a4865bd946d9a59e634dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffe2a69e-85", "ovs_interfaceid": "ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.789555] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eb422e-4539-20e4-2897-fd127a683316, 'name': SearchDatastore_Task, 'duration_secs': 0.010042} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.789859] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.790105] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1011.790352] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.790503] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.790686] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1011.790955] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff324a1d-fc0b-4fb6-a7b4-6bfc0cbb510a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.799454] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1011.799578] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1011.800342] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c61e868-3f95-4ec0-8648-aa24aaf798a2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.810028] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1011.810028] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b169af-1f8a-ff5c-3ea0-3c7b6d98871e" [ 1011.810028] env[63197]: _type = "Task" [ 1011.810028] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.818691] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b169af-1f8a-ff5c-3ea0-3c7b6d98871e, 'name': SearchDatastore_Task, 'duration_secs': 0.008224} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.819756] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c55aed77-1a03-43b4-8417-019c565e6a1b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.824799] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1011.824799] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5242d7a6-c99e-4541-ed07-7dea6bf59061" [ 1011.824799] env[63197]: _type = "Task" [ 1011.824799] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.832232] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5242d7a6-c99e-4541-ed07-7dea6bf59061, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.949516] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.043950] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364502, 'name': ReconfigVM_Task, 'duration_secs': 0.285281} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.044104] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Reconfigured VM instance instance-0000005a to attach disk [datastore1] f39f3e80-0d96-4024-aec8-16e20e9bf13f/f39f3e80-0d96-4024-aec8-16e20e9bf13f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.044714] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4912eee1-f6ba-4c6d-8ce2-96588426ef52 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.050621] env[63197]: DEBUG nova.scheduler.client.report [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.054866] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1012.054866] env[63197]: value = "task-1364503" [ 1012.054866] env[63197]: _type = "Task" [ 1012.054866] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.066432] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364503, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.179343] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Releasing lock "refresh_cache-d8919322-1fa7-4fc7-9380-a8b7d283b050" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.179750] env[63197]: DEBUG nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Instance network_info: |[{"id": "ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52", "address": "fa:16:3e:ee:64:45", "network": {"id": "72ef7714-d486-4e2b-a3e7-c52ede2f7dd6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.56", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ba0c482e322a4865bd946d9a59e634dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffe2a69e-85", "ovs_interfaceid": "ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1012.180261] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:64:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.187879] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Creating folder: Project (1ece2f02eca2440986f0f5c720fb7d1e). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.188231] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a9fe638b-bd49-41f0-9f92-8eecd5cdf6bd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.200601] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Created folder: Project (1ece2f02eca2440986f0f5c720fb7d1e) in parent group-v290286. [ 1012.200791] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Creating folder: Instances. Parent ref: group-v290393. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1012.201033] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b70829a8-ae7f-467b-8ced-5b49bb609d4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.209158] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Created folder: Instances in parent group-v290393. [ 1012.209353] env[63197]: DEBUG oslo.service.loopingcall [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.209538] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1012.209729] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-75ac7d95-f099-42a1-9305-5de5b92cbde9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.228253] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.228253] env[63197]: value = "task-1364506" [ 1012.228253] env[63197]: _type = "Task" [ 1012.228253] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.236267] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364506, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.312327] env[63197]: DEBUG nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1012.336475] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5242d7a6-c99e-4541-ed07-7dea6bf59061, 'name': SearchDatastore_Task, 'duration_secs': 0.009915} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.336828] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.337198] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4/cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1012.337455] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-131a92f8-b11b-4082-9109-2ba74ca49bdd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.343823] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.344081] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.344251] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.344502] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.344693] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.345031] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.345205] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.345429] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.345622] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.345791] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.346041] env[63197]: DEBUG nova.virt.hardware [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.346972] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5325f6-1356-4c9d-9c41-6ef1793e4242 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.351116] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1012.351116] env[63197]: value = "task-1364507" [ 1012.351116] env[63197]: _type = "Task" [ 1012.351116] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.358232] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc3d151-ac8b-430e-a865-005fa6fa0221 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.365174] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364507, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.448711] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.502394] env[63197]: DEBUG nova.compute.manager [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Received event network-changed-ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.502615] env[63197]: DEBUG nova.compute.manager [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Refreshing instance network info cache due to event network-changed-ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1012.502912] env[63197]: DEBUG oslo_concurrency.lockutils [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] Acquiring lock "refresh_cache-d8919322-1fa7-4fc7-9380-a8b7d283b050" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.503234] env[63197]: DEBUG oslo_concurrency.lockutils [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] Acquired lock "refresh_cache-d8919322-1fa7-4fc7-9380-a8b7d283b050" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.503479] env[63197]: DEBUG nova.network.neutron [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Refreshing network info cache for port ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1012.556461] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.557025] env[63197]: DEBUG nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.560269] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.744s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.560483] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.563380] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.775s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.563380] env[63197]: DEBUG nova.objects.instance [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'resources' on Instance uuid 8e5ef1b0-7532-498a-84c2-189274a36c50 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.575414] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364503, 'name': Rename_Task, 'duration_secs': 0.149644} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.575889] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1012.575967] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b505292-492d-4359-b57c-a42ff6294bbb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.582761] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1012.582761] env[63197]: value = "task-1364508" [ 1012.582761] env[63197]: _type = "Task" [ 1012.582761] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.592712] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364508, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.593799] env[63197]: INFO nova.scheduler.client.report [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted allocations for instance 64b944ea-2397-45eb-a0ed-9f57263db998 [ 1012.656095] env[63197]: DEBUG nova.network.neutron [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Successfully updated port: 9161195e-f24b-453c-b121-a6a965441223 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.741141] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364506, 'name': CreateVM_Task, 'duration_secs': 0.29783} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.741319] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1012.742036] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.742291] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.742681] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.742984] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-278f1cbb-5fa6-42e1-a9c0-5481d859e7fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.748857] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1012.748857] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f898de-4fbb-5b36-9d6a-21b59b4ac5ea" [ 1012.748857] env[63197]: _type = "Task" [ 1012.748857] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.757609] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f898de-4fbb-5b36-9d6a-21b59b4ac5ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.860921] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364507, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.948758] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.070089] env[63197]: DEBUG nova.compute.utils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.075282] env[63197]: DEBUG nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.075401] env[63197]: DEBUG nova.network.neutron [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1013.092500] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364508, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.103566] env[63197]: DEBUG oslo_concurrency.lockutils [None req-790e496d-a207-425b-bb3d-387dc23a8df2 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "64b944ea-2397-45eb-a0ed-9f57263db998" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.345s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.116942] env[63197]: DEBUG nova.policy [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b50be4b75a94b4481c9c65ea1e4e9ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bb89fb32d8c4726a9a3104d68ce560a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.154143] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "refresh_cache-a77fe81f-fa44-4639-acbb-bf093b5d4f6c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.154211] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired lock "refresh_cache-a77fe81f-fa44-4639-acbb-bf093b5d4f6c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.154367] env[63197]: DEBUG nova.network.neutron [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1013.230302] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bc431e-1c61-483f-b390-86bb14e26ba3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.237943] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b413775e-2efb-415f-814d-f371a093f6b5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.266726] env[63197]: DEBUG nova.network.neutron [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Updated VIF entry in instance network info cache for port ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1013.267180] env[63197]: DEBUG nova.network.neutron [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Updating instance_info_cache with network_info: [{"id": "ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52", "address": "fa:16:3e:ee:64:45", "network": {"id": "72ef7714-d486-4e2b-a3e7-c52ede2f7dd6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.56", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ba0c482e322a4865bd946d9a59e634dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapffe2a69e-85", "ovs_interfaceid": "ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.272231] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8581db6e-9819-43c2-ad10-50c77479cfbb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.282212] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f898de-4fbb-5b36-9d6a-21b59b4ac5ea, 'name': SearchDatastore_Task, 'duration_secs': 0.049951} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.284307] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.284574] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.284847] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.285071] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.285321] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.285871] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6fe5a0b-0540-4332-b750-5e72d90cf6d0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.288465] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6ac863-bff2-4051-a90c-c56d6ebd963a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.302451] env[63197]: DEBUG nova.compute.provider_tree [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.305383] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.305628] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.306363] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7454af5a-9598-442b-b67d-f26374c25397 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.311659] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1013.311659] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c4c6a4-6d98-7408-1ee6-1c7d7b5e3967" [ 1013.311659] env[63197]: _type = "Task" [ 1013.311659] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.319905] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c4c6a4-6d98-7408-1ee6-1c7d7b5e3967, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.363474] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364507, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532611} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.363933] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4/cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.364486] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.364905] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6088e9f4-b8c5-467f-a871-00373b7f3799 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.371939] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1013.371939] env[63197]: value = "task-1364509" [ 1013.371939] env[63197]: _type = "Task" [ 1013.371939] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.377142] env[63197]: DEBUG nova.network.neutron [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Successfully created port: 8e81ce3c-4538-4717-a05b-29e5b48aed2d {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.383116] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.449612] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.575773] env[63197]: DEBUG nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.592544] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364508, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.691838] env[63197]: DEBUG nova.network.neutron [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1013.777573] env[63197]: DEBUG oslo_concurrency.lockutils [req-c0be36a2-3dd7-4c32-8d6b-ebe05bdbe78f req-f351ec98-7f40-438a-851d-1bbd44ebac63 service nova] Releasing lock "refresh_cache-d8919322-1fa7-4fc7-9380-a8b7d283b050" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.805892] env[63197]: DEBUG nova.scheduler.client.report [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.823371] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c4c6a4-6d98-7408-1ee6-1c7d7b5e3967, 'name': SearchDatastore_Task, 'duration_secs': 0.018267} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.824422] env[63197]: DEBUG nova.network.neutron [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Updating instance_info_cache with network_info: [{"id": "9161195e-f24b-453c-b121-a6a965441223", "address": "fa:16:3e:f8:b4:20", "network": {"id": "72ef7714-d486-4e2b-a3e7-c52ede2f7dd6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ba0c482e322a4865bd946d9a59e634dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9161195e-f2", "ovs_interfaceid": "9161195e-f24b-453c-b121-a6a965441223", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.826926] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d6c174b-c95e-42ce-a84c-01b33de13d5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.832996] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1013.832996] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c8e69e-0724-cb59-7d1b-d3ffec459fad" [ 1013.832996] env[63197]: _type = "Task" [ 1013.832996] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.841455] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c8e69e-0724-cb59-7d1b-d3ffec459fad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.880834] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.239124} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.881095] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.881848] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47cc2c3-2dc3-48b3-81be-4368dd418f6c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.904102] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4/cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.904380] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2a01e32-448a-4b9e-a819-35e2a5fc8755 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.923890] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1013.923890] env[63197]: value = "task-1364510" [ 1013.923890] env[63197]: _type = "Task" [ 1013.923890] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.932733] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364510, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.949200] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.094656] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364508, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.312834] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.750s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.315592] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.892s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.328128] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Releasing lock "refresh_cache-a77fe81f-fa44-4639-acbb-bf093b5d4f6c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.328657] env[63197]: DEBUG nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Instance network_info: |[{"id": "9161195e-f24b-453c-b121-a6a965441223", "address": "fa:16:3e:f8:b4:20", "network": {"id": "72ef7714-d486-4e2b-a3e7-c52ede2f7dd6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ba0c482e322a4865bd946d9a59e634dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9161195e-f2", "ovs_interfaceid": "9161195e-f24b-453c-b121-a6a965441223", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1014.328792] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:b4:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9161195e-f24b-453c-b121-a6a965441223', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.336338] env[63197]: DEBUG oslo.service.loopingcall [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.336625] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1014.339802] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ee615ad-ce5c-4b69-a32e-fd831a7e669a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.355797] env[63197]: INFO nova.scheduler.client.report [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted allocations for instance 8e5ef1b0-7532-498a-84c2-189274a36c50 [ 1014.365477] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c8e69e-0724-cb59-7d1b-d3ffec459fad, 'name': SearchDatastore_Task, 'duration_secs': 0.011169} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.366731] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.367029] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] d8919322-1fa7-4fc7-9380-a8b7d283b050/d8919322-1fa7-4fc7-9380-a8b7d283b050.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.367274] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.367274] env[63197]: value = "task-1364511" [ 1014.367274] env[63197]: _type = "Task" [ 1014.367274] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.367483] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2fc3597-a7a4-4ea4-8953-369eb8dfe15e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.377630] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364511, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.379495] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1014.379495] env[63197]: value = "task-1364512" [ 1014.379495] env[63197]: _type = "Task" [ 1014.379495] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.386560] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.434079] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364510, 'name': ReconfigVM_Task, 'duration_secs': 0.292106} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.434407] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Reconfigured VM instance instance-0000005b to attach disk [datastore1] cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4/cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.435105] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93583f47-bd1e-4c51-9bb8-47c1865021a6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.442238] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1014.442238] env[63197]: value = "task-1364513" [ 1014.442238] env[63197]: _type = "Task" [ 1014.442238] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.453847] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.457145] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364513, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.529569] env[63197]: DEBUG nova.compute.manager [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Received event network-vif-plugged-9161195e-f24b-453c-b121-a6a965441223 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.529847] env[63197]: DEBUG oslo_concurrency.lockutils [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] Acquiring lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.530093] env[63197]: DEBUG oslo_concurrency.lockutils [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.530298] env[63197]: DEBUG oslo_concurrency.lockutils [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.530498] env[63197]: DEBUG nova.compute.manager [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] No waiting events found dispatching network-vif-plugged-9161195e-f24b-453c-b121-a6a965441223 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.530718] env[63197]: WARNING nova.compute.manager [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Received unexpected event network-vif-plugged-9161195e-f24b-453c-b121-a6a965441223 for instance with vm_state building and task_state spawning. [ 1014.530922] env[63197]: DEBUG nova.compute.manager [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Received event network-changed-9161195e-f24b-453c-b121-a6a965441223 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.531115] env[63197]: DEBUG nova.compute.manager [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Refreshing instance network info cache due to event network-changed-9161195e-f24b-453c-b121-a6a965441223. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.531337] env[63197]: DEBUG oslo_concurrency.lockutils [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] Acquiring lock "refresh_cache-a77fe81f-fa44-4639-acbb-bf093b5d4f6c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.531537] env[63197]: DEBUG oslo_concurrency.lockutils [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] Acquired lock "refresh_cache-a77fe81f-fa44-4639-acbb-bf093b5d4f6c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.531672] env[63197]: DEBUG nova.network.neutron [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Refreshing network info cache for port 9161195e-f24b-453c-b121-a6a965441223 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1014.592794] env[63197]: DEBUG nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.600782] env[63197]: DEBUG oslo_vmware.api [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364508, 'name': PowerOnVM_Task, 'duration_secs': 1.568206} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.601214] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1014.601448] env[63197]: INFO nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Took 9.09 seconds to spawn the instance on the hypervisor. [ 1014.601736] env[63197]: DEBUG nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.602554] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc97edc-7779-400c-a723-0434cb6013d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.625570] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.625958] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.626256] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.626494] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.626688] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.626860] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.627099] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.627271] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.627442] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.627611] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.627786] env[63197]: DEBUG nova.virt.hardware [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.628754] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca689111-8b9b-47dd-85f8-ecf146203a90 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.638104] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b3ae47-d367-48cf-8da7-1626c14ee946 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.785933] env[63197]: DEBUG nova.compute.manager [req-d404dca7-6c66-4ed0-8b02-5011a5fae34e req-68f20e42-4c11-4554-8b19-50714a859801 service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Received event network-vif-plugged-8e81ce3c-4538-4717-a05b-29e5b48aed2d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.786388] env[63197]: DEBUG oslo_concurrency.lockutils [req-d404dca7-6c66-4ed0-8b02-5011a5fae34e req-68f20e42-4c11-4554-8b19-50714a859801 service nova] Acquiring lock "7f9ad29e-9cb9-4575-8dce-081c02767af5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.787014] env[63197]: DEBUG oslo_concurrency.lockutils [req-d404dca7-6c66-4ed0-8b02-5011a5fae34e req-68f20e42-4c11-4554-8b19-50714a859801 service nova] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.787285] env[63197]: DEBUG oslo_concurrency.lockutils [req-d404dca7-6c66-4ed0-8b02-5011a5fae34e req-68f20e42-4c11-4554-8b19-50714a859801 service nova] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.787538] env[63197]: DEBUG nova.compute.manager [req-d404dca7-6c66-4ed0-8b02-5011a5fae34e req-68f20e42-4c11-4554-8b19-50714a859801 service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] No waiting events found dispatching network-vif-plugged-8e81ce3c-4538-4717-a05b-29e5b48aed2d {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.787839] env[63197]: WARNING nova.compute.manager [req-d404dca7-6c66-4ed0-8b02-5011a5fae34e req-68f20e42-4c11-4554-8b19-50714a859801 service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Received unexpected event network-vif-plugged-8e81ce3c-4538-4717-a05b-29e5b48aed2d for instance with vm_state building and task_state spawning. [ 1014.873296] env[63197]: DEBUG oslo_concurrency.lockutils [None req-905dc0ae-4caf-4c9a-8fbe-0fa3a92070f2 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "8e5ef1b0-7532-498a-84c2-189274a36c50" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.216s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.880121] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364511, 'name': CreateVM_Task, 'duration_secs': 0.414449} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.883249] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.884135] env[63197]: DEBUG nova.network.neutron [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Successfully updated port: 8e81ce3c-4538-4717-a05b-29e5b48aed2d {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1014.885731] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.885953] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.886359] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.887261] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cfe9e00-ff03-4e35-91b0-a37536f4f369 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.892345] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364512, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488326} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.893760] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] d8919322-1fa7-4fc7-9380-a8b7d283b050/d8919322-1fa7-4fc7-9380-a8b7d283b050.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1014.893976] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1014.894243] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9786020-a62b-449b-b488-dee82ee413d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.897702] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1014.897702] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527586af-a26d-8cac-60dc-84e90d71c29f" [ 1014.897702] env[63197]: _type = "Task" [ 1014.897702] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.902728] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1014.902728] env[63197]: value = "task-1364514" [ 1014.902728] env[63197]: _type = "Task" [ 1014.902728] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.908801] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527586af-a26d-8cac-60dc-84e90d71c29f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.913190] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364514, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.952402] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.957341] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364513, 'name': Rename_Task, 'duration_secs': 0.143174} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.957654] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.957916] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e78a8edd-7ea8-43c9-af9c-2b0ab715694b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.964529] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1014.964529] env[63197]: value = "task-1364515" [ 1014.964529] env[63197]: _type = "Task" [ 1014.964529] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.972251] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.125786] env[63197]: INFO nova.compute.manager [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Took 13.94 seconds to build instance. [ 1015.243609] env[63197]: DEBUG nova.network.neutron [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Updated VIF entry in instance network info cache for port 9161195e-f24b-453c-b121-a6a965441223. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1015.244023] env[63197]: DEBUG nova.network.neutron [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Updating instance_info_cache with network_info: [{"id": "9161195e-f24b-453c-b121-a6a965441223", "address": "fa:16:3e:f8:b4:20", "network": {"id": "72ef7714-d486-4e2b-a3e7-c52ede2f7dd6", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "ba0c482e322a4865bd946d9a59e634dc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9161195e-f2", "ovs_interfaceid": "9161195e-f24b-453c-b121-a6a965441223", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 638ef9c9-253b-4958-a660-6c1801408a51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 4672f595-e1f6-4400-b5a1-065598584980 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 180e0da7-f7ee-4fcd-be95-c2bf679278d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance f39f3e80-0d96-4024-aec8-16e20e9bf13f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance d8919322-1fa7-4fc7-9380-a8b7d283b050 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance a77fe81f-fa44-4639-acbb-bf093b5d4f6c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 7f9ad29e-9cb9-4575-8dce-081c02767af5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1015.355268] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2240MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1015.389429] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "refresh_cache-7f9ad29e-9cb9-4575-8dce-081c02767af5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.389562] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "refresh_cache-7f9ad29e-9cb9-4575-8dce-081c02767af5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.389710] env[63197]: DEBUG nova.network.neutron [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1015.411125] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527586af-a26d-8cac-60dc-84e90d71c29f, 'name': SearchDatastore_Task, 'duration_secs': 0.010751} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.413975] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.414249] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.414486] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.414636] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.414816] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.415972] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1c35f12-a4de-411a-9388-b0549dfba0fb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.420366] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364514, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090736} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.421032] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.421723] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b03abd-3bc9-41df-ba65-e56a5e14a694 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.445957] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] d8919322-1fa7-4fc7-9380-a8b7d283b050/d8919322-1fa7-4fc7-9380-a8b7d283b050.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.446144] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.446255] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1015.449706] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb001b6f-3208-438d-b97a-32889eb82061 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.464299] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4f2bc23-33b1-4dcd-9e2d-ecd88982ebdb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.475736] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1015.475736] env[63197]: value = "task-1364516" [ 1015.475736] env[63197]: _type = "Task" [ 1015.475736] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.477130] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1015.477130] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5220cef9-6237-2f46-4deb-a5e400fc92c9" [ 1015.477130] env[63197]: _type = "Task" [ 1015.477130] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.489313] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.489607] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364515, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.502723] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.503359] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5220cef9-6237-2f46-4deb-a5e400fc92c9, 'name': SearchDatastore_Task, 'duration_secs': 0.020475} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.503865] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a2f5bdf-160c-47a0-aaff-8736d5c09cc2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.511315] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1015.511315] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5222097b-ad66-d50c-af77-dc6249286a05" [ 1015.511315] env[63197]: _type = "Task" [ 1015.511315] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.519572] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5222097b-ad66-d50c-af77-dc6249286a05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.533482] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1da2c49-bae5-473b-95a8-e66fe1905500 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.540980] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd11d3d-515b-49d9-9ec2-77729a30df41 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.575021] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb30c8e3-f58e-424a-82a3-e8ad05a7d3eb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.582507] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394471a8-9666-4356-8148-7b4e58b77e53 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.595798] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.628135] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0579448f-e12e-49cc-8f93-bc54fa602fc9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.478s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.747038] env[63197]: DEBUG oslo_concurrency.lockutils [req-43e9676b-2795-4627-b0ba-682134539c9f req-60bc591a-dfe6-4932-8b88-343d08124c93 service nova] Releasing lock "refresh_cache-a77fe81f-fa44-4639-acbb-bf093b5d4f6c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.891953] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.892256] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.892478] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.892732] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.892934] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.897518] env[63197]: INFO nova.compute.manager [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Terminating instance [ 1015.899482] env[63197]: DEBUG nova.compute.manager [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1015.899753] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.900733] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9af6a34-68a8-4fbe-84c9-86ddef40ae29 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.908495] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.908894] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb65003b-d305-41a6-8c35-b973335e9716 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.914806] env[63197]: DEBUG oslo_vmware.api [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1015.914806] env[63197]: value = "task-1364517" [ 1015.914806] env[63197]: _type = "Task" [ 1015.914806] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.924415] env[63197]: DEBUG oslo_vmware.api [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364517, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.934835] env[63197]: DEBUG nova.network.neutron [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1015.962130] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.978849] env[63197]: DEBUG oslo_vmware.api [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364515, 'name': PowerOnVM_Task, 'duration_secs': 0.527886} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.982142] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.982370] env[63197]: INFO nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Took 8.22 seconds to spawn the instance on the hypervisor. [ 1015.982559] env[63197]: DEBUG nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.983358] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4627ab7-6a92-4eef-a850-cf70e16c7d95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.991355] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.021680] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5222097b-ad66-d50c-af77-dc6249286a05, 'name': SearchDatastore_Task, 'duration_secs': 0.009632} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.021935] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.022226] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] a77fe81f-fa44-4639-acbb-bf093b5d4f6c/a77fe81f-fa44-4639-acbb-bf093b5d4f6c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1016.023595] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f00d3c9-26fc-485a-b7d0-a10341798325 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.029395] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1016.029395] env[63197]: value = "task-1364518" [ 1016.029395] env[63197]: _type = "Task" [ 1016.029395] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.038241] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364518, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.098591] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.292647] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "638ef9c9-253b-4958-a660-6c1801408a51" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.292992] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.331846] env[63197]: DEBUG nova.network.neutron [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Updating instance_info_cache with network_info: [{"id": "8e81ce3c-4538-4717-a05b-29e5b48aed2d", "address": "fa:16:3e:04:14:db", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e81ce3c-45", "ovs_interfaceid": "8e81ce3c-4538-4717-a05b-29e5b48aed2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.424880] env[63197]: DEBUG oslo_vmware.api [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364517, 'name': PowerOffVM_Task, 'duration_secs': 0.298481} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.425098] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1016.425260] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1016.425543] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7832010-d47d-4670-9d9b-85027eb17b9b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.460577] env[63197]: DEBUG oslo_vmware.api [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364499, 'name': ReconfigVM_Task, 'duration_secs': 5.796503} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.460893] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.461208] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Reconfigured VM to detach interface {{(pid=63197) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1016.490257] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364516, 'name': ReconfigVM_Task, 'duration_secs': 0.774488} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.490780] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Reconfigured VM instance instance-0000005c to attach disk [datastore1] d8919322-1fa7-4fc7-9380-a8b7d283b050/d8919322-1fa7-4fc7-9380-a8b7d283b050.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.491224] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ca327b7-bba0-437e-8c2f-da0e02e5ee0b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.496800] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1016.497019] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1016.497203] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleting the datastore file [datastore1] f39f3e80-0d96-4024-aec8-16e20e9bf13f {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1016.497489] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9359a129-43ec-4a69-b9b8-4a8b081c9276 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.507189] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1016.507189] env[63197]: value = "task-1364520" [ 1016.507189] env[63197]: _type = "Task" [ 1016.507189] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.512917] env[63197]: DEBUG oslo_vmware.api [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1016.512917] env[63197]: value = "task-1364521" [ 1016.512917] env[63197]: _type = "Task" [ 1016.512917] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.513707] env[63197]: INFO nova.compute.manager [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Took 14.59 seconds to build instance. [ 1016.521636] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364520, 'name': Rename_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.527228] env[63197]: DEBUG oslo_vmware.api [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364521, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.540462] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364518, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.604140] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1016.604346] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.289s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.796043] env[63197]: INFO nova.compute.manager [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Detaching volume 6665b511-2b56-4af2-8d68-620525b85aa5 [ 1016.812110] env[63197]: DEBUG nova.compute.manager [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Received event network-changed-8e81ce3c-4538-4717-a05b-29e5b48aed2d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.812317] env[63197]: DEBUG nova.compute.manager [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Refreshing instance network info cache due to event network-changed-8e81ce3c-4538-4717-a05b-29e5b48aed2d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1016.812514] env[63197]: DEBUG oslo_concurrency.lockutils [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] Acquiring lock "refresh_cache-7f9ad29e-9cb9-4575-8dce-081c02767af5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.835401] env[63197]: INFO nova.virt.block_device [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Attempting to driver detach volume 6665b511-2b56-4af2-8d68-620525b85aa5 from mountpoint /dev/sdb [ 1016.835520] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1016.835705] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290363', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'name': 'volume-6665b511-2b56-4af2-8d68-620525b85aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '638ef9c9-253b-4958-a660-6c1801408a51', 'attached_at': '', 'detached_at': '', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'serial': '6665b511-2b56-4af2-8d68-620525b85aa5'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1016.836247] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "refresh_cache-7f9ad29e-9cb9-4575-8dce-081c02767af5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.836535] env[63197]: DEBUG nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Instance network_info: |[{"id": "8e81ce3c-4538-4717-a05b-29e5b48aed2d", "address": "fa:16:3e:04:14:db", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e81ce3c-45", "ovs_interfaceid": "8e81ce3c-4538-4717-a05b-29e5b48aed2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1016.837516] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7d6ec0-1044-43ff-bff4-6305963c020c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.840263] env[63197]: DEBUG oslo_concurrency.lockutils [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] Acquired lock "refresh_cache-7f9ad29e-9cb9-4575-8dce-081c02767af5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.840449] env[63197]: DEBUG nova.network.neutron [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Refreshing network info cache for port 8e81ce3c-4538-4717-a05b-29e5b48aed2d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1016.842156] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:14:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fd77ecbc-aaaf-45f4-ae8f-977d90e4052f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e81ce3c-4538-4717-a05b-29e5b48aed2d', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1016.850450] env[63197]: DEBUG oslo.service.loopingcall [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.851632] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1016.852311] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-92c3dd73-1c2c-4de3-a572-4a48d6a9c6b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.884164] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd995f8-f92c-4702-9cc5-79d1d03e4f1b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.890008] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1016.890008] env[63197]: value = "task-1364522" [ 1016.890008] env[63197]: _type = "Task" [ 1016.890008] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.897929] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46c0ec3-d77b-4c8c-86ed-ae180faf5454 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.903173] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364522, 'name': CreateVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.921403] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d9f96a-3ca4-4dae-ace3-1ef1efffa78a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.936440] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] The volume has not been displaced from its original location: [datastore1] volume-6665b511-2b56-4af2-8d68-620525b85aa5/volume-6665b511-2b56-4af2-8d68-620525b85aa5.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1016.941682] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1016.941975] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e88a54c-91b1-435f-819b-96885420e0d0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.959273] env[63197]: DEBUG oslo_vmware.api [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1016.959273] env[63197]: value = "task-1364523" [ 1016.959273] env[63197]: _type = "Task" [ 1016.959273] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.967203] env[63197]: DEBUG oslo_vmware.api [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364523, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.017276] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364520, 'name': Rename_Task, 'duration_secs': 0.209022} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.020380] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.020865] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d66a2aa5-ab9a-4965-8b4c-354d7bff3d6d tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.104s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.021091] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d01eb2a1-b576-40c9-b792-87a3fe46a280 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.027349] env[63197]: DEBUG oslo_vmware.api [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364521, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278838} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.028476] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1017.028669] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1017.028852] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1017.029040] env[63197]: INFO nova.compute.manager [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1017.029283] env[63197]: DEBUG oslo.service.loopingcall [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.029531] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1017.029531] env[63197]: value = "task-1364524" [ 1017.029531] env[63197]: _type = "Task" [ 1017.029531] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.029724] env[63197]: DEBUG nova.compute.manager [-] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1017.029818] env[63197]: DEBUG nova.network.neutron [-] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1017.041295] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364518, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634275} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.044302] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] a77fe81f-fa44-4639-acbb-bf093b5d4f6c/a77fe81f-fa44-4639-acbb-bf093b5d4f6c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1017.044520] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.044775] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364524, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.045043] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2cf19f60-c036-4639-8e9e-db372a00e97e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.051610] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1017.051610] env[63197]: value = "task-1364525" [ 1017.051610] env[63197]: _type = "Task" [ 1017.051610] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.060272] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364525, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.400196] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364522, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.471880] env[63197]: DEBUG oslo_vmware.api [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364523, 'name': ReconfigVM_Task, 'duration_secs': 0.264824} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.472224] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1017.476842] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5647e3e1-b55e-4420-baee-e1cf25c8cafb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.495474] env[63197]: DEBUG oslo_vmware.api [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1017.495474] env[63197]: value = "task-1364526" [ 1017.495474] env[63197]: _type = "Task" [ 1017.495474] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.503522] env[63197]: DEBUG oslo_vmware.api [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364526, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.544519] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364524, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.563585] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364525, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075909} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.563804] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.564644] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4367830-c0e8-4d6e-b9b3-f6437ae18c95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.589219] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] a77fe81f-fa44-4639-acbb-bf093b5d4f6c/a77fe81f-fa44-4639-acbb-bf093b5d4f6c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.589219] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-abfd5249-e8a3-480c-bf5f-a306c9577540 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.608238] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1017.608238] env[63197]: value = "task-1364527" [ 1017.608238] env[63197]: _type = "Task" [ 1017.608238] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.616516] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364527, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.617341] env[63197]: DEBUG nova.network.neutron [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Updated VIF entry in instance network info cache for port 8e81ce3c-4538-4717-a05b-29e5b48aed2d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1017.617658] env[63197]: DEBUG nova.network.neutron [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Updating instance_info_cache with network_info: [{"id": "8e81ce3c-4538-4717-a05b-29e5b48aed2d", "address": "fa:16:3e:04:14:db", "network": {"id": "012b92ad-57e2-4bf4-a0d5-0df4b2f20221", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2008940973-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bb89fb32d8c4726a9a3104d68ce560a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fd77ecbc-aaaf-45f4-ae8f-977d90e4052f", "external-id": "nsx-vlan-transportzone-171", "segmentation_id": 171, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e81ce3c-45", "ovs_interfaceid": "8e81ce3c-4538-4717-a05b-29e5b48aed2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.749853] env[63197]: DEBUG nova.network.neutron [-] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.896200] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.896425] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.896972] env[63197]: DEBUG nova.network.neutron [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.907533] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364522, 'name': CreateVM_Task, 'duration_secs': 0.531014} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.907693] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1017.909039] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.909039] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.909039] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1017.909209] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d096611-d701-4859-a939-cee9042b6641 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.914371] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1017.914371] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52309d67-9f52-87c2-4fe5-464b45156497" [ 1017.914371] env[63197]: _type = "Task" [ 1017.914371] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.922667] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52309d67-9f52-87c2-4fe5-464b45156497, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.005717] env[63197]: DEBUG oslo_vmware.api [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364526, 'name': ReconfigVM_Task, 'duration_secs': 0.138625} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.006081] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290363', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'name': 'volume-6665b511-2b56-4af2-8d68-620525b85aa5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '638ef9c9-253b-4958-a660-6c1801408a51', 'attached_at': '', 'detached_at': '', 'volume_id': '6665b511-2b56-4af2-8d68-620525b85aa5', 'serial': '6665b511-2b56-4af2-8d68-620525b85aa5'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1018.041591] env[63197]: DEBUG oslo_vmware.api [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364524, 'name': PowerOnVM_Task, 'duration_secs': 0.684412} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.041856] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.042065] env[63197]: INFO nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1018.042243] env[63197]: DEBUG nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.043099] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad56e25c-9435-4986-a806-2b9cea00cccb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.121463] env[63197]: DEBUG oslo_concurrency.lockutils [req-a1249216-9ae8-4091-9158-f02b2d723dd9 req-812b6dd9-5129-4685-b7b0-81cf37583b6f service nova] Releasing lock "refresh_cache-7f9ad29e-9cb9-4575-8dce-081c02767af5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.121997] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.249642] env[63197]: DEBUG oslo_concurrency.lockutils [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.249979] env[63197]: DEBUG oslo_concurrency.lockutils [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.250255] env[63197]: DEBUG nova.compute.manager [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.251282] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e4ab70-9686-402e-8e04-912df0050ec2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.254657] env[63197]: INFO nova.compute.manager [-] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Took 1.22 seconds to deallocate network for instance. [ 1018.263139] env[63197]: DEBUG nova.compute.manager [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63197) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1018.263139] env[63197]: DEBUG nova.objects.instance [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lazy-loading 'flavor' on Instance uuid cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.426452] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52309d67-9f52-87c2-4fe5-464b45156497, 'name': SearchDatastore_Task, 'duration_secs': 0.009715} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.427393] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.427851] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.428259] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.428498] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.428860] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.429326] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c90caecd-2372-4fdc-84b0-4de2d779eb5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.439814] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.439967] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.442944] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06c5d409-ae73-49e9-85d0-f2dce6968ebf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.448413] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1018.448413] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522a6485-aaaf-5b65-d531-30bfcdbf3615" [ 1018.448413] env[63197]: _type = "Task" [ 1018.448413] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.456479] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522a6485-aaaf-5b65-d531-30bfcdbf3615, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.558454] env[63197]: DEBUG nova.objects.instance [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'flavor' on Instance uuid 638ef9c9-253b-4958-a660-6c1801408a51 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.564015] env[63197]: INFO nova.compute.manager [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Took 16.24 seconds to build instance. [ 1018.620766] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364527, 'name': ReconfigVM_Task, 'duration_secs': 0.655884} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.621054] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Reconfigured VM instance instance-0000005d to attach disk [datastore1] a77fe81f-fa44-4639-acbb-bf093b5d4f6c/a77fe81f-fa44-4639-acbb-bf093b5d4f6c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.621682] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6eb30f42-4a4e-4310-bc5d-2404df5b37ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.628417] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1018.628417] env[63197]: value = "task-1364528" [ 1018.628417] env[63197]: _type = "Task" [ 1018.628417] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.637721] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364528, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.709634] env[63197]: INFO nova.network.neutron [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Port 44d19552-d824-4f52-93fa-832e41750693 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1018.710051] env[63197]: DEBUG nova.network.neutron [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.765943] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.766283] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.766533] env[63197]: DEBUG nova.objects.instance [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'resources' on Instance uuid f39f3e80-0d96-4024-aec8-16e20e9bf13f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.771801] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.772080] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff0ddd4d-03a8-4646-a4df-4ed50b76420c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.780858] env[63197]: DEBUG oslo_vmware.api [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1018.780858] env[63197]: value = "task-1364529" [ 1018.780858] env[63197]: _type = "Task" [ 1018.780858] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.789586] env[63197]: DEBUG oslo_vmware.api [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.867819] env[63197]: DEBUG nova.compute.manager [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Received event network-vif-deleted-0284be0f-c3dd-4db2-8a6b-10ec63920553 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.868614] env[63197]: DEBUG nova.compute.manager [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.868614] env[63197]: DEBUG nova.compute.manager [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing instance network info cache due to event network-changed-6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1018.868614] env[63197]: DEBUG oslo_concurrency.lockutils [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] Acquiring lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.960357] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522a6485-aaaf-5b65-d531-30bfcdbf3615, 'name': SearchDatastore_Task, 'duration_secs': 0.025931} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.961954] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-561eca93-3423-4625-9aba-c175ae92c4f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.967641] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1018.967641] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a8a906-3a3f-2cb2-3651-5743e2418bcb" [ 1018.967641] env[63197]: _type = "Task" [ 1018.967641] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.978195] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a8a906-3a3f-2cb2-3651-5743e2418bcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.066504] env[63197]: DEBUG oslo_concurrency.lockutils [None req-b6c52acb-02bb-458a-8933-38ef05af62c1 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.751s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.140977] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364528, 'name': Rename_Task, 'duration_secs': 0.220307} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.141371] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.141984] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41afd2ac-c946-498b-a7ea-84341363cae9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.147495] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1019.147495] env[63197]: value = "task-1364530" [ 1019.147495] env[63197]: _type = "Task" [ 1019.147495] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.155463] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364530, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.213016] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.215347] env[63197]: DEBUG oslo_concurrency.lockutils [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] Acquired lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.215639] env[63197]: DEBUG nova.network.neutron [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Refreshing network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1019.290638] env[63197]: DEBUG oslo_vmware.api [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364529, 'name': PowerOffVM_Task, 'duration_secs': 0.211373} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.293145] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.293335] env[63197]: DEBUG nova.compute.manager [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.294242] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d07ab39-9de3-4234-b729-fb44b56c70c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.305495] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-44d19552-d824-4f52-93fa-832e41750693" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.305719] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-44d19552-d824-4f52-93fa-832e41750693" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.306085] env[63197]: DEBUG nova.objects.instance [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'flavor' on Instance uuid 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.381771] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb873dc-3132-4440-92b9-e0cb0071c92a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.388972] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c081ada3-06bc-4a19-b367-ff500dc4a515 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.417734] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f79df5-b73e-47ab-a172-f3681de7addd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.424798] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d960c662-9197-4eaa-a41a-a9fdd9d10afa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.439103] env[63197]: DEBUG nova.compute.provider_tree [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.476979] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a8a906-3a3f-2cb2-3651-5743e2418bcb, 'name': SearchDatastore_Task, 'duration_secs': 0.021045} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.477248] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.477502] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 7f9ad29e-9cb9-4575-8dce-081c02767af5/7f9ad29e-9cb9-4575-8dce-081c02767af5.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.477740] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9268349-d081-4162-b6d4-9ea56048cf35 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.485096] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1019.485096] env[63197]: value = "task-1364531" [ 1019.485096] env[63197]: _type = "Task" [ 1019.485096] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.492512] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.568879] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ead66728-26a7-4a45-baca-ed75a3872c1a tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.276s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.658431] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364530, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.718884] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ca38cb3e-1704-4aa7-880e-6777b6f58560 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-4672f595-e1f6-4400-b5a1-065598584980-44d19552-d824-4f52-93fa-832e41750693" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.907s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.811867] env[63197]: DEBUG oslo_concurrency.lockutils [None req-beeeb68d-2dc5-4fa0-b808-b8d9f7126b0f tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.940225] env[63197]: DEBUG nova.objects.instance [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'pci_requests' on Instance uuid 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.943096] env[63197]: DEBUG nova.scheduler.client.report [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.997881] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469223} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.998363] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 7f9ad29e-9cb9-4575-8dce-081c02767af5/7f9ad29e-9cb9-4575-8dce-081c02767af5.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1019.998728] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1019.999196] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6092df5-e783-40ec-8d44-3634cd5444ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.006287] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1020.006287] env[63197]: value = "task-1364532" [ 1020.006287] env[63197]: _type = "Task" [ 1020.006287] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.013714] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.060303] env[63197]: DEBUG nova.network.neutron [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updated VIF entry in instance network info cache for port 6085d1e1-acf6-4471-8a56-e050285a562d. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1020.060627] env[63197]: DEBUG nova.network.neutron [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [{"id": "6085d1e1-acf6-4471-8a56-e050285a562d", "address": "fa:16:3e:9f:3a:fe", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6085d1e1-ac", "ovs_interfaceid": "6085d1e1-acf6-4471-8a56-e050285a562d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.159792] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364530, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.447738] env[63197]: DEBUG nova.objects.base [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Object Instance<1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f> lazy-loaded attributes: flavor,pci_requests {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1020.447738] env[63197]: DEBUG nova.network.neutron [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1020.450368] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.473323] env[63197]: INFO nova.scheduler.client.report [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted allocations for instance f39f3e80-0d96-4024-aec8-16e20e9bf13f [ 1020.516629] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062837} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.517537] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.518336] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a12674-48c3-43bb-87ee-5c50d0ba03a7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.541792] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 7f9ad29e-9cb9-4575-8dce-081c02767af5/7f9ad29e-9cb9-4575-8dce-081c02767af5.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.542350] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-861cb030-9c9b-4031-b9fd-99f34d89240f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.560114] env[63197]: DEBUG nova.policy [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '417ac7ca823c43e2b7cad63607e1a946', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8db3cc951174f6192ff954ff4d704de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1020.563098] env[63197]: DEBUG oslo_concurrency.lockutils [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] Releasing lock "refresh_cache-4672f595-e1f6-4400-b5a1-065598584980" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.563368] env[63197]: DEBUG nova.compute.manager [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.563627] env[63197]: DEBUG nova.compute.manager [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing instance network info cache due to event network-changed-6de5e286-38ff-452d-b567-9b6c956831d1. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.563747] env[63197]: DEBUG oslo_concurrency.lockutils [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.563903] env[63197]: DEBUG oslo_concurrency.lockutils [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.564117] env[63197]: DEBUG nova.network.neutron [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing network info cache for port 6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.566445] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1020.566445] env[63197]: value = "task-1364533" [ 1020.566445] env[63197]: _type = "Task" [ 1020.566445] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.574657] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364533, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.659735] env[63197]: DEBUG oslo_vmware.api [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364530, 'name': PowerOnVM_Task, 'duration_secs': 1.263843} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.660138] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.660293] env[63197]: INFO nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Took 8.35 seconds to spawn the instance on the hypervisor. [ 1020.660375] env[63197]: DEBUG nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.661262] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646c6e6b-cb33-4b04-81c5-a11081c874d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.698486] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "638ef9c9-253b-4958-a660-6c1801408a51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.698787] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.699034] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "638ef9c9-253b-4958-a660-6c1801408a51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.699262] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.699441] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.705075] env[63197]: INFO nova.compute.manager [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Terminating instance [ 1020.707419] env[63197]: DEBUG nova.compute.manager [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1020.707618] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1020.708424] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1fd9380-0c2e-4887-af56-8e9171e1098b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.716473] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.716709] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04bcb38f-d946-48e9-a4ec-f6962391cd89 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.722614] env[63197]: DEBUG oslo_vmware.api [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1020.722614] env[63197]: value = "task-1364534" [ 1020.722614] env[63197]: _type = "Task" [ 1020.722614] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.730390] env[63197]: DEBUG oslo_vmware.api [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364534, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.982268] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cf47af3e-09aa-42ea-99a0-ca556a5bf48b tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "f39f3e80-0d96-4024-aec8-16e20e9bf13f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.090s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.078067] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364533, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.178882] env[63197]: INFO nova.compute.manager [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Took 18.66 seconds to build instance. [ 1021.235263] env[63197]: DEBUG oslo_vmware.api [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364534, 'name': PowerOffVM_Task, 'duration_secs': 0.210791} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.235501] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1021.235674] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.235927] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18e9aed6-d0af-4ca5-98dc-9c91781251b9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.271758] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.272063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.272290] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.272483] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.272656] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.275017] env[63197]: INFO nova.compute.manager [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Terminating instance [ 1021.277079] env[63197]: DEBUG nova.compute.manager [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.277340] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.278216] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae4da5f-2209-40f4-99eb-1454efb7c0b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.285744] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1021.286014] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa1a3e39-198a-4539-9be7-834849f87d2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.295316] env[63197]: DEBUG nova.network.neutron [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updated VIF entry in instance network info cache for port 6de5e286-38ff-452d-b567-9b6c956831d1. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1021.295691] env[63197]: DEBUG nova.network.neutron [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.298839] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.299066] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.299253] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleting the datastore file [datastore2] 638ef9c9-253b-4958-a660-6c1801408a51 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.299500] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7304e4db-8d9b-4bde-be71-59462e7e6702 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.305792] env[63197]: DEBUG oslo_vmware.api [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1021.305792] env[63197]: value = "task-1364537" [ 1021.305792] env[63197]: _type = "Task" [ 1021.305792] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.316856] env[63197]: DEBUG oslo_vmware.api [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.352588] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1021.352874] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1021.353263] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleting the datastore file [datastore1] cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.353357] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a63394f1-a0db-46e9-a1db-7e3c747d03b1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.360184] env[63197]: DEBUG oslo_vmware.api [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for the task: (returnval){ [ 1021.360184] env[63197]: value = "task-1364538" [ 1021.360184] env[63197]: _type = "Task" [ 1021.360184] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.368715] env[63197]: DEBUG oslo_vmware.api [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364538, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.577060] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364533, 'name': ReconfigVM_Task, 'duration_secs': 0.856298} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.577415] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 7f9ad29e-9cb9-4575-8dce-081c02767af5/7f9ad29e-9cb9-4575-8dce-081c02767af5.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.577949] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-589f84c5-edab-4f81-a75e-edec1f105d5e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.584381] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1021.584381] env[63197]: value = "task-1364539" [ 1021.584381] env[63197]: _type = "Task" [ 1021.584381] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.592823] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364539, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.681027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c115a5ca-36f3-4b98-91a3-9a25050eaff6 tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.173s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.798411] env[63197]: DEBUG oslo_concurrency.lockutils [req-c9a8c031-884f-494c-a45a-7432b2675b61 req-b88b6f31-66b0-4ed2-9aae-2f235f0c5fff service nova] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.816989] env[63197]: DEBUG oslo_vmware.api [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159519} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.817290] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.817497] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.817671] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.817863] env[63197]: INFO nova.compute.manager [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1021.818141] env[63197]: DEBUG oslo.service.loopingcall [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.818346] env[63197]: DEBUG nova.compute.manager [-] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1021.818444] env[63197]: DEBUG nova.network.neutron [-] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.869573] env[63197]: DEBUG oslo_vmware.api [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Task: {'id': task-1364538, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151447} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.869866] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1021.870016] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1021.870209] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1021.870384] env[63197]: INFO nova.compute.manager [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1021.870623] env[63197]: DEBUG oslo.service.loopingcall [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1021.870842] env[63197]: DEBUG nova.compute.manager [-] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1021.870928] env[63197]: DEBUG nova.network.neutron [-] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1021.987784] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "af31d3b0-52ed-408c-8110-c4241e204424" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.988026] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.094990] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364539, 'name': Rename_Task, 'duration_secs': 0.278954} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.095405] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.095721] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0dbb7d01-b117-4bec-beea-0c1f2d69f885 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.106108] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1022.106108] env[63197]: value = "task-1364540" [ 1022.106108] env[63197]: _type = "Task" [ 1022.106108] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.110375] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364540, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.247472] env[63197]: DEBUG nova.compute.manager [req-b514ed98-444f-4f01-ab34-6fd90750361e req-16f97a7c-5793-480a-a3e9-fbd976f7f6d9 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-vif-plugged-44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.247690] env[63197]: DEBUG oslo_concurrency.lockutils [req-b514ed98-444f-4f01-ab34-6fd90750361e req-16f97a7c-5793-480a-a3e9-fbd976f7f6d9 service nova] Acquiring lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.247904] env[63197]: DEBUG oslo_concurrency.lockutils [req-b514ed98-444f-4f01-ab34-6fd90750361e req-16f97a7c-5793-480a-a3e9-fbd976f7f6d9 service nova] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.248091] env[63197]: DEBUG oslo_concurrency.lockutils [req-b514ed98-444f-4f01-ab34-6fd90750361e req-16f97a7c-5793-480a-a3e9-fbd976f7f6d9 service nova] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.248262] env[63197]: DEBUG nova.compute.manager [req-b514ed98-444f-4f01-ab34-6fd90750361e req-16f97a7c-5793-480a-a3e9-fbd976f7f6d9 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] No waiting events found dispatching network-vif-plugged-44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.248715] env[63197]: WARNING nova.compute.manager [req-b514ed98-444f-4f01-ab34-6fd90750361e req-16f97a7c-5793-480a-a3e9-fbd976f7f6d9 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received unexpected event network-vif-plugged-44d19552-d824-4f52-93fa-832e41750693 for instance with vm_state active and task_state None. [ 1022.296556] env[63197]: DEBUG nova.compute.manager [req-4db24447-6eae-4717-8335-740188d9fd6e req-b5ad02b0-97d6-4ce2-8085-c2eebac74cb3 service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Received event network-vif-deleted-e927a6cd-692a-4cce-b2f6-8d321333e34f {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.296556] env[63197]: INFO nova.compute.manager [req-4db24447-6eae-4717-8335-740188d9fd6e req-b5ad02b0-97d6-4ce2-8085-c2eebac74cb3 service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Neutron deleted interface e927a6cd-692a-4cce-b2f6-8d321333e34f; detaching it from the instance and deleting it from the info cache [ 1022.296556] env[63197]: DEBUG nova.network.neutron [req-4db24447-6eae-4717-8335-740188d9fd6e req-b5ad02b0-97d6-4ce2-8085-c2eebac74cb3 service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.377297] env[63197]: DEBUG nova.network.neutron [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Successfully updated port: 44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.491024] env[63197]: DEBUG nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1022.612164] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364540, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.720203] env[63197]: DEBUG nova.compute.manager [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.721192] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba7fa6d-db89-4ef4-9e02-eda73d05335c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.770034] env[63197]: DEBUG nova.network.neutron [-] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.800749] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c3b703c-1202-48f6-b36a-79feb809f796 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.808918] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b842461f-32d9-4de4-b895-18c17994a44d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.822647] env[63197]: DEBUG nova.network.neutron [-] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.838193] env[63197]: DEBUG nova.compute.manager [req-4db24447-6eae-4717-8335-740188d9fd6e req-b5ad02b0-97d6-4ce2-8085-c2eebac74cb3 service nova] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Detach interface failed, port_id=e927a6cd-692a-4cce-b2f6-8d321333e34f, reason: Instance 638ef9c9-253b-4958-a660-6c1801408a51 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1022.879908] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.880208] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.880421] env[63197]: DEBUG nova.network.neutron [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1023.016041] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.016041] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.017769] env[63197]: INFO nova.compute.claims [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.112601] env[63197]: DEBUG oslo_vmware.api [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364540, 'name': PowerOnVM_Task, 'duration_secs': 0.953974} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.112982] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.112982] env[63197]: INFO nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Took 8.52 seconds to spawn the instance on the hypervisor. [ 1023.113192] env[63197]: DEBUG nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.113999] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff51d62-bf84-4132-8b60-e185699195ae {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.232665] env[63197]: INFO nova.compute.manager [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] instance snapshotting [ 1023.236245] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b91e8f-3458-4100-8b12-137afba5fed4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.259030] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f78082-1d88-4b58-a855-010b95cb99d7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.272151] env[63197]: INFO nova.compute.manager [-] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Took 1.45 seconds to deallocate network for instance. [ 1023.325417] env[63197]: INFO nova.compute.manager [-] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Took 1.45 seconds to deallocate network for instance. [ 1023.414578] env[63197]: WARNING nova.network.neutron [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] 7ac08e03-1288-4a87-9a9c-257ba778c6d6 already exists in list: networks containing: ['7ac08e03-1288-4a87-9a9c-257ba778c6d6']. ignoring it [ 1023.635763] env[63197]: INFO nova.compute.manager [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Took 19.40 seconds to build instance. [ 1023.691298] env[63197]: DEBUG nova.network.neutron [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "44d19552-d824-4f52-93fa-832e41750693", "address": "fa:16:3e:35:a8:2e", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d19552-d8", "ovs_interfaceid": "44d19552-d824-4f52-93fa-832e41750693", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.768495] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1023.769154] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-af9f88a8-9793-4af2-9611-d139bd690a9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.776376] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1023.776376] env[63197]: value = "task-1364541" [ 1023.776376] env[63197]: _type = "Task" [ 1023.776376] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.777242] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.785604] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364541, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.831071] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.138332] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7a8467c1-da32-42b6-99ef-626ce81bfb25 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.912s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.140034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ceed03-ea74-4419-9064-31f5e707ab5e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.146701] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e6f6bbc-7fd6-4dca-9da8-f444be1c7590 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.177054] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447073e1-3ff4-4e19-bafa-efa3faf214ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.184464] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88db66ff-9eb5-447b-95ba-4912fd3f054e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.199111] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.199762] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.199930] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.200373] env[63197]: DEBUG nova.compute.provider_tree [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.202113] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb54c5a2-39a2-4388-9730-d41a004ad985 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.218850] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.219112] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.219275] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.219467] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.219623] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.219811] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.220047] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.220250] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.220420] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.220587] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.220783] env[63197]: DEBUG nova.virt.hardware [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.227327] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Reconfiguring VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1024.228956] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08e86481-de45-4f83-a2f7-8aea5ea27d7c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.245445] env[63197]: DEBUG oslo_vmware.api [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1024.245445] env[63197]: value = "task-1364542" [ 1024.245445] env[63197]: _type = "Task" [ 1024.245445] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.253813] env[63197]: DEBUG oslo_vmware.api [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364542, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.286038] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364541, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.336986] env[63197]: DEBUG nova.compute.manager [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Received event network-vif-deleted-7a9f024c-f13b-437a-8b6c-ea3c7752044a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.337248] env[63197]: DEBUG nova.compute.manager [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-changed-44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.337432] env[63197]: DEBUG nova.compute.manager [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing instance network info cache due to event network-changed-44d19552-d824-4f52-93fa-832e41750693. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1024.337660] env[63197]: DEBUG oslo_concurrency.lockutils [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.337821] env[63197]: DEBUG oslo_concurrency.lockutils [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.337984] env[63197]: DEBUG nova.network.neutron [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Refreshing network info cache for port 44d19552-d824-4f52-93fa-832e41750693 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1024.707075] env[63197]: DEBUG nova.scheduler.client.report [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1024.756656] env[63197]: DEBUG oslo_vmware.api [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364542, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.786916] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364541, 'name': CreateSnapshot_Task, 'duration_secs': 0.868655} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.787489] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1024.787972] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79c5118-b4bf-474a-bfdf-2ba470b873ba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.085761] env[63197]: DEBUG nova.network.neutron [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updated VIF entry in instance network info cache for port 44d19552-d824-4f52-93fa-832e41750693. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1025.086424] env[63197]: DEBUG nova.network.neutron [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "44d19552-d824-4f52-93fa-832e41750693", "address": "fa:16:3e:35:a8:2e", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44d19552-d8", "ovs_interfaceid": "44d19552-d824-4f52-93fa-832e41750693", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.211983] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.196s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.212561] env[63197]: DEBUG nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1025.215238] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.438s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.215465] env[63197]: DEBUG nova.objects.instance [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'resources' on Instance uuid 638ef9c9-253b-4958-a660-6c1801408a51 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.258067] env[63197]: DEBUG oslo_vmware.api [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364542, 'name': ReconfigVM_Task, 'duration_secs': 0.573561} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.258609] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.258767] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Reconfigured VM to attach interface {{(pid=63197) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1025.306369] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1025.307043] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-73ed785f-1cbe-48e8-ac6f-f04b6133b549 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.316149] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1025.316149] env[63197]: value = "task-1364543" [ 1025.316149] env[63197]: _type = "Task" [ 1025.316149] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.324241] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364543, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.380540] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "7f9ad29e-9cb9-4575-8dce-081c02767af5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.380923] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.381238] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "7f9ad29e-9cb9-4575-8dce-081c02767af5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.381507] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.381762] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.384511] env[63197]: INFO nova.compute.manager [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Terminating instance [ 1025.386591] env[63197]: DEBUG nova.compute.manager [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.386804] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.387704] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deaee4aa-2878-4188-b514-9097bd40dac9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.396666] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.397053] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a91ec1f-827a-4d6d-b581-91fec7685581 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.405306] env[63197]: DEBUG oslo_vmware.api [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1025.405306] env[63197]: value = "task-1364544" [ 1025.405306] env[63197]: _type = "Task" [ 1025.405306] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.414739] env[63197]: DEBUG oslo_vmware.api [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.589926] env[63197]: DEBUG oslo_concurrency.lockutils [req-c8d0c1e0-3d8d-4a3d-88e3-ae794cd2bd7a req-cb6e3022-3f74-468b-b8ba-3b5d8ff68964 service nova] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.719144] env[63197]: DEBUG nova.compute.utils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1025.723696] env[63197]: DEBUG nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1025.723875] env[63197]: DEBUG nova.network.neutron [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1025.762814] env[63197]: DEBUG nova.policy [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a8e87b64fe46d7ab7e570d57611119', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ceb1016d6d34bff8880dca42d495377', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1025.765300] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bcc17b67-7397-4d5e-a321-3ee94a5de425 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-44d19552-d824-4f52-93fa-832e41750693" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.459s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.827708] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364543, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.853586] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770492ee-9fa8-478d-a14f-9dd9c062afcf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.863263] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a846d1-784d-4cf4-b474-3b22ed44353e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.914202] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730e680a-e5af-4c52-8418-11ec17bed111 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.925465] env[63197]: DEBUG oslo_vmware.api [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364544, 'name': PowerOffVM_Task, 'duration_secs': 0.308925} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.928884] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.929216] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1025.929683] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-52328492-fe26-4537-b93c-16afd87549fa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.933657] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad4bbd3-1da0-4d98-97d0-ebc68e001174 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.955769] env[63197]: DEBUG nova.compute.provider_tree [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.011738] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.012100] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.012401] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleting the datastore file [datastore2] 7f9ad29e-9cb9-4575-8dce-081c02767af5 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.012800] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-52ba3d07-3ab4-4c3f-8dcb-0623991ded76 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.019411] env[63197]: DEBUG oslo_vmware.api [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for the task: (returnval){ [ 1026.019411] env[63197]: value = "task-1364546" [ 1026.019411] env[63197]: _type = "Task" [ 1026.019411] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.028533] env[63197]: DEBUG oslo_vmware.api [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364546, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.053059] env[63197]: DEBUG nova.network.neutron [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Successfully created port: 9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1026.224846] env[63197]: DEBUG nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1026.338280] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364543, 'name': CloneVM_Task, 'duration_secs': 0.856331} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.338518] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Created linked-clone VM from snapshot [ 1026.341244] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8427f36-178b-4b91-ac60-08c1faeb4be2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.347385] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Uploading image 6faea748-1bf7-4227-849d-e15dfcb02ddf {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1026.376309] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1026.376309] env[63197]: value = "vm-290399" [ 1026.376309] env[63197]: _type = "VirtualMachine" [ 1026.376309] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1026.376637] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-31fd9e5d-faed-40ec-bc94-4fa3ea64fe9b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.383142] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lease: (returnval){ [ 1026.383142] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eeb72b-511d-a159-a650-9ba9ed63bb9b" [ 1026.383142] env[63197]: _type = "HttpNfcLease" [ 1026.383142] env[63197]: } obtained for exporting VM: (result){ [ 1026.383142] env[63197]: value = "vm-290399" [ 1026.383142] env[63197]: _type = "VirtualMachine" [ 1026.383142] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1026.383701] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the lease: (returnval){ [ 1026.383701] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eeb72b-511d-a159-a650-9ba9ed63bb9b" [ 1026.383701] env[63197]: _type = "HttpNfcLease" [ 1026.383701] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1026.389921] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1026.389921] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eeb72b-511d-a159-a650-9ba9ed63bb9b" [ 1026.389921] env[63197]: _type = "HttpNfcLease" [ 1026.389921] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1026.460151] env[63197]: DEBUG nova.scheduler.client.report [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.531541] env[63197]: DEBUG oslo_vmware.api [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Task: {'id': task-1364546, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139731} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.531834] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.532132] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.532333] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.532513] env[63197]: INFO nova.compute.manager [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1026.532774] env[63197]: DEBUG oslo.service.loopingcall [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.533107] env[63197]: DEBUG nova.compute.manager [-] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.533211] env[63197]: DEBUG nova.network.neutron [-] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.892547] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1026.892547] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eeb72b-511d-a159-a650-9ba9ed63bb9b" [ 1026.892547] env[63197]: _type = "HttpNfcLease" [ 1026.892547] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1026.893014] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1026.893014] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52eeb72b-511d-a159-a650-9ba9ed63bb9b" [ 1026.893014] env[63197]: _type = "HttpNfcLease" [ 1026.893014] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1026.894025] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa6f173-c492-4343-aa71-d665602c5fa5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.903244] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52616849-3284-94a8-a3bc-543425f4f9aa/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1026.903423] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52616849-3284-94a8-a3bc-543425f4f9aa/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1026.965923] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.751s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.968115] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.137s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.968420] env[63197]: DEBUG nova.objects.instance [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lazy-loading 'resources' on Instance uuid cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.973384] env[63197]: DEBUG nova.compute.manager [req-d09f70bf-7e99-4a99-a7c6-17cf9d17f336 req-679b0487-0645-4cf5-ac5d-e06b2934e345 service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Received event network-vif-deleted-8e81ce3c-4538-4717-a05b-29e5b48aed2d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.973662] env[63197]: INFO nova.compute.manager [req-d09f70bf-7e99-4a99-a7c6-17cf9d17f336 req-679b0487-0645-4cf5-ac5d-e06b2934e345 service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Neutron deleted interface 8e81ce3c-4538-4717-a05b-29e5b48aed2d; detaching it from the instance and deleting it from the info cache [ 1026.973753] env[63197]: DEBUG nova.network.neutron [req-d09f70bf-7e99-4a99-a7c6-17cf9d17f336 req-679b0487-0645-4cf5-ac5d-e06b2934e345 service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.992334] env[63197]: INFO nova.scheduler.client.report [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted allocations for instance 638ef9c9-253b-4958-a660-6c1801408a51 [ 1027.030915] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0145527e-e890-4a1c-95fc-6c2ee7a20136 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.052563] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "interface-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-44d19552-d824-4f52-93fa-832e41750693" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.052842] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-44d19552-d824-4f52-93fa-832e41750693" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.236100] env[63197]: DEBUG nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.257292] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.257594] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.257777] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.257979] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.258161] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.258326] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.258549] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.258717] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.258909] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.259098] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.259293] env[63197]: DEBUG nova.virt.hardware [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.260233] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3b310e-b6ab-49ec-95cb-1e2d2aa56fbb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.268889] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6062fd-38d3-477c-a62d-e1fef43afffe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.467139] env[63197]: DEBUG nova.network.neutron [-] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.476259] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-24425b71-9b48-4603-a28a-be20d14dee1d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.491398] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e1f3af-bc48-41de-9a78-0e5077c10711 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.511932] env[63197]: DEBUG oslo_concurrency.lockutils [None req-54b5d54c-b0fe-4744-a35a-d1d7bf97094c tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "638ef9c9-253b-4958-a660-6c1801408a51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.812s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.535778] env[63197]: DEBUG nova.compute.manager [req-d09f70bf-7e99-4a99-a7c6-17cf9d17f336 req-679b0487-0645-4cf5-ac5d-e06b2934e345 service nova] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Detach interface failed, port_id=8e81ce3c-4538-4717-a05b-29e5b48aed2d, reason: Instance 7f9ad29e-9cb9-4575-8dce-081c02767af5 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1027.560034] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.560034] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.560034] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1358cfc-fa1d-441a-8aaf-fa7cdc257e01 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.582833] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce6ff64-b69d-4eaf-844e-a02a1b52e2ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.611629] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Reconfiguring VM to detach interface {{(pid=63197) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1027.615953] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-001b0227-2dd7-47c2-9769-0f7811ea6832 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.637628] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1027.637628] env[63197]: value = "task-1364548" [ 1027.637628] env[63197]: _type = "Task" [ 1027.637628] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.647319] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.706210] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd488ff8-a751-4c28-aae5-e9e8cc312304 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.711270] env[63197]: DEBUG nova.network.neutron [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Successfully updated port: 9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.717347] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f7827f-4ac6-4fab-a769-52f5929d3415 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.754754] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1435cad9-3d08-4b54-b9b3-329e87a89ebc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.763389] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b9294d-0475-42a3-a05d-0ccb7ed5d9ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.779930] env[63197]: DEBUG nova.compute.provider_tree [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.972034] env[63197]: INFO nova.compute.manager [-] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Took 1.44 seconds to deallocate network for instance. [ 1028.151496] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.216646] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-af31d3b0-52ed-408c-8110-c4241e204424" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.216825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-af31d3b0-52ed-408c-8110-c4241e204424" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.217086] env[63197]: DEBUG nova.network.neutron [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.283335] env[63197]: DEBUG nova.scheduler.client.report [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.479060] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.649618] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.751915] env[63197]: DEBUG nova.network.neutron [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1028.788903] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.821s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.791555] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.313s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.791852] env[63197]: DEBUG nova.objects.instance [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lazy-loading 'resources' on Instance uuid 7f9ad29e-9cb9-4575-8dce-081c02767af5 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.814885] env[63197]: INFO nova.scheduler.client.report [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Deleted allocations for instance cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4 [ 1028.969556] env[63197]: DEBUG nova.network.neutron [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Updating instance_info_cache with network_info: [{"id": "9b9071c4-4e71-4761-bd7c-ab1367d3f9e6", "address": "fa:16:3e:94:3d:aa", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b9071c4-4e", "ovs_interfaceid": "9b9071c4-4e71-4761-bd7c-ab1367d3f9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.001247] env[63197]: DEBUG nova.compute.manager [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Received event network-vif-plugged-9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.001555] env[63197]: DEBUG oslo_concurrency.lockutils [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] Acquiring lock "af31d3b0-52ed-408c-8110-c4241e204424-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.001695] env[63197]: DEBUG oslo_concurrency.lockutils [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] Lock "af31d3b0-52ed-408c-8110-c4241e204424-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.001888] env[63197]: DEBUG oslo_concurrency.lockutils [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] Lock "af31d3b0-52ed-408c-8110-c4241e204424-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.002089] env[63197]: DEBUG nova.compute.manager [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] No waiting events found dispatching network-vif-plugged-9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1029.002266] env[63197]: WARNING nova.compute.manager [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Received unexpected event network-vif-plugged-9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 for instance with vm_state building and task_state spawning. [ 1029.002430] env[63197]: DEBUG nova.compute.manager [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Received event network-changed-9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.002606] env[63197]: DEBUG nova.compute.manager [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Refreshing instance network info cache due to event network-changed-9b9071c4-4e71-4761-bd7c-ab1367d3f9e6. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.002834] env[63197]: DEBUG oslo_concurrency.lockutils [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] Acquiring lock "refresh_cache-af31d3b0-52ed-408c-8110-c4241e204424" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.149866] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.326370] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8840efa4-015e-4952-924c-d2fd2faa8ca0 tempest-ServersTestJSON-982444055 tempest-ServersTestJSON-982444055-project-member] Lock "cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.054s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.395911] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421f92e8-680f-4e69-9b1f-5b6a62f8973d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.403439] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854166e5-4026-478f-9dae-593480fff554 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.435482] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de26984f-fcdc-490f-a8fd-bc737265e0ec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.443036] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c850b5-fd8b-4793-8144-220732c7f8e2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.456646] env[63197]: DEBUG nova.compute.provider_tree [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.472253] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-af31d3b0-52ed-408c-8110-c4241e204424" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.472550] env[63197]: DEBUG nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Instance network_info: |[{"id": "9b9071c4-4e71-4761-bd7c-ab1367d3f9e6", "address": "fa:16:3e:94:3d:aa", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b9071c4-4e", "ovs_interfaceid": "9b9071c4-4e71-4761-bd7c-ab1367d3f9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.472909] env[63197]: DEBUG oslo_concurrency.lockutils [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] Acquired lock "refresh_cache-af31d3b0-52ed-408c-8110-c4241e204424" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.473132] env[63197]: DEBUG nova.network.neutron [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Refreshing network info cache for port 9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.474683] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:3d:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9b9071c4-4e71-4761-bd7c-ab1367d3f9e6', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.481899] env[63197]: DEBUG oslo.service.loopingcall [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.482891] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1029.483159] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a72583e8-464a-4195-90ea-2037b881c932 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.503768] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.503768] env[63197]: value = "task-1364549" [ 1029.503768] env[63197]: _type = "Task" [ 1029.503768] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.512111] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364549, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.649718] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.813832] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.814238] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.960157] env[63197]: DEBUG nova.scheduler.client.report [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.014609] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364549, 'name': CreateVM_Task, 'duration_secs': 0.340968} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.014878] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1030.015555] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.015728] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.016063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1030.016376] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-451f5eb4-d388-4351-9206-625c07ec753c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.021790] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1030.021790] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277a69b-63d1-b7ae-a50a-208f78fff8ff" [ 1030.021790] env[63197]: _type = "Task" [ 1030.021790] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.033472] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277a69b-63d1-b7ae-a50a-208f78fff8ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.149803] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.203377] env[63197]: DEBUG nova.network.neutron [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Updated VIF entry in instance network info cache for port 9b9071c4-4e71-4761-bd7c-ab1367d3f9e6. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1030.203800] env[63197]: DEBUG nova.network.neutron [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Updating instance_info_cache with network_info: [{"id": "9b9071c4-4e71-4761-bd7c-ab1367d3f9e6", "address": "fa:16:3e:94:3d:aa", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9b9071c4-4e", "ovs_interfaceid": "9b9071c4-4e71-4761-bd7c-ab1367d3f9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.317146] env[63197]: DEBUG nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1030.467462] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.674s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.500093] env[63197]: INFO nova.scheduler.client.report [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Deleted allocations for instance 7f9ad29e-9cb9-4575-8dce-081c02767af5 [ 1030.533314] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5277a69b-63d1-b7ae-a50a-208f78fff8ff, 'name': SearchDatastore_Task, 'duration_secs': 0.012618} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.533684] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.533949] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1030.534354] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.534448] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.534579] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.534879] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aad4d99d-20cf-4768-84f0-78cb9b281189 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.543365] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.543598] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1030.544383] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c51cb6ee-25d1-447a-b599-507714bf7953 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.550678] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1030.550678] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bf88c4-7294-8e91-fd17-72427337bdad" [ 1030.550678] env[63197]: _type = "Task" [ 1030.550678] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.559302] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bf88c4-7294-8e91-fd17-72427337bdad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.650315] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.706827] env[63197]: DEBUG oslo_concurrency.lockutils [req-88875a15-4304-4209-9d18-981c8e0361c4 req-49bd219b-4184-401d-bb28-1906ad5cf76f service nova] Releasing lock "refresh_cache-af31d3b0-52ed-408c-8110-c4241e204424" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.842227] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.842891] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.846786] env[63197]: INFO nova.compute.claims [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1031.010880] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2d5c460b-bc55-4945-a79f-092bdc84cfa5 tempest-ServerDiskConfigTestJSON-1969395996 tempest-ServerDiskConfigTestJSON-1969395996-project-member] Lock "7f9ad29e-9cb9-4575-8dce-081c02767af5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.630s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.061522] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bf88c4-7294-8e91-fd17-72427337bdad, 'name': SearchDatastore_Task, 'duration_secs': 0.009332} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.062361] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8913a1c6-6bdd-4a88-9047-011d3954d03e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.067595] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1031.067595] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527f238e-8334-aaff-be79-26615d96f416" [ 1031.067595] env[63197]: _type = "Task" [ 1031.067595] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.075364] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527f238e-8334-aaff-be79-26615d96f416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.151216] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.578701] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]527f238e-8334-aaff-be79-26615d96f416, 'name': SearchDatastore_Task, 'duration_secs': 0.008753} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.578963] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.579259] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] af31d3b0-52ed-408c-8110-c4241e204424/af31d3b0-52ed-408c-8110-c4241e204424.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1031.579527] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1ced641-339a-461d-a3b9-f62dae914b1d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.586362] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1031.586362] env[63197]: value = "task-1364550" [ 1031.586362] env[63197]: _type = "Task" [ 1031.586362] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.595792] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364550, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.652338] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.005304] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ba8aba-50b8-4fc0-8fe0-d318fbc03710 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.014101] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c41b02-cc35-434e-a3aa-d10dce64b544 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.047484] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e7ad063-64b4-48a5-896a-781321c072c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.056105] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a482a82d-86b0-4095-a9e8-a4268a2384e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.072797] env[63197]: DEBUG nova.compute.provider_tree [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.096286] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364550, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494553} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.096591] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] af31d3b0-52ed-408c-8110-c4241e204424/af31d3b0-52ed-408c-8110-c4241e204424.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1032.096823] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.097172] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff1946fb-d049-4d7b-8c7e-153664280085 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.103481] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1032.103481] env[63197]: value = "task-1364551" [ 1032.103481] env[63197]: _type = "Task" [ 1032.103481] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.112519] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364551, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.151536] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.576536] env[63197]: DEBUG nova.scheduler.client.report [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.613567] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364551, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060338} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.613642] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.614440] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b497e3-e7f1-42cc-9ddd-163f554a5401 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.636612] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] af31d3b0-52ed-408c-8110-c4241e204424/af31d3b0-52ed-408c-8110-c4241e204424.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.636926] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8f2bf3f-7598-4dc4-b889-253263c478b6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.661269] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.662666] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1032.662666] env[63197]: value = "task-1364552" [ 1032.662666] env[63197]: _type = "Task" [ 1032.662666] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.670437] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364552, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.090999] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.242s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.090999] env[63197]: DEBUG nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1033.161263] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.171955] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364552, 'name': ReconfigVM_Task, 'duration_secs': 0.284452} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.172319] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Reconfigured VM instance instance-0000005f to attach disk [datastore2] af31d3b0-52ed-408c-8110-c4241e204424/af31d3b0-52ed-408c-8110-c4241e204424.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.173036] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ca287ff-021e-4d6d-b3cf-74a00c12697b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.179444] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1033.179444] env[63197]: value = "task-1364553" [ 1033.179444] env[63197]: _type = "Task" [ 1033.179444] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.187940] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364553, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.597075] env[63197]: DEBUG nova.compute.utils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1033.597400] env[63197]: DEBUG nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1033.597793] env[63197]: DEBUG nova.network.neutron [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1033.643497] env[63197]: DEBUG nova.policy [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '514b7795ecdc49069e7f3cab700a0ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '527721557235413e99c4a2eaa086486c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1033.661338] env[63197]: DEBUG oslo_vmware.api [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364548, 'name': ReconfigVM_Task, 'duration_secs': 5.795254} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.661542] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.661591] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Reconfigured VM to detach interface {{(pid=63197) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1033.692284] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364553, 'name': Rename_Task, 'duration_secs': 0.13905} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.692575] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.692830] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c2f356a-2ffe-49e9-9de1-fb1a95f215e9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.700497] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1033.700497] env[63197]: value = "task-1364554" [ 1033.700497] env[63197]: _type = "Task" [ 1033.700497] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.708222] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.102565] env[63197]: DEBUG nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1034.138549] env[63197]: DEBUG nova.network.neutron [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Successfully created port: 8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1034.216929] env[63197]: DEBUG oslo_vmware.api [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364554, 'name': PowerOnVM_Task, 'duration_secs': 0.467108} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.218447] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.218777] env[63197]: INFO nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Took 6.98 seconds to spawn the instance on the hypervisor. [ 1034.219116] env[63197]: DEBUG nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.220695] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806e2b73-4afa-45c6-8253-5af2f0ba379d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.741266] env[63197]: INFO nova.compute.manager [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Took 11.75 seconds to build instance. [ 1034.968690] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52616849-3284-94a8-a3bc-543425f4f9aa/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1034.969769] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4fa621-d414-416c-a498-00477ca03f93 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.977116] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52616849-3284-94a8-a3bc-543425f4f9aa/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1034.977293] env[63197]: ERROR oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52616849-3284-94a8-a3bc-543425f4f9aa/disk-0.vmdk due to incomplete transfer. [ 1034.977525] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c27aee1c-2cd5-4866-98b5-38d851832b2d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.984571] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52616849-3284-94a8-a3bc-543425f4f9aa/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1034.984824] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Uploaded image 6faea748-1bf7-4227-849d-e15dfcb02ddf to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1034.987435] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1034.987697] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-45a03dd8-9a12-47f7-a878-a73ea91a9a8b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.993543] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1034.993543] env[63197]: value = "task-1364555" [ 1034.993543] env[63197]: _type = "Task" [ 1034.993543] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.001102] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364555, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.118518] env[63197]: DEBUG nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1035.140443] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.140748] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.140917] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.141132] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.141291] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.141441] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.141681] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.141866] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.142053] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.142223] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.142402] env[63197]: DEBUG nova.virt.hardware [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.143347] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2f4a51-9408-4576-8cb5-e2c115ea71b7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.152043] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2111f92-918d-4736-aa68-3d342fdbde3f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.217307] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "1b234498-8fe0-475a-9cad-71d0be79307c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.217560] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "1b234498-8fe0-475a-9cad-71d0be79307c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.221542] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.221721] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquired lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.221902] env[63197]: DEBUG nova.network.neutron [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1035.243190] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3e557206-6237-4bbd-8c6c-57901dfd4423 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.255s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.407741] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "2ab00a06-545f-4674-b7e3-37354c7460a4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.407978] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "2ab00a06-545f-4674-b7e3-37354c7460a4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.504421] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364555, 'name': Destroy_Task, 'duration_secs': 0.310019} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.504729] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Destroyed the VM [ 1035.504997] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1035.505386] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c79a39b8-b90d-4901-a346-dc86869f77ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.511755] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1035.511755] env[63197]: value = "task-1364556" [ 1035.511755] env[63197]: _type = "Task" [ 1035.511755] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.519911] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364556, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.561563] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.562084] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.562356] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.562568] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.562743] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.564942] env[63197]: INFO nova.compute.manager [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Terminating instance [ 1035.566800] env[63197]: DEBUG nova.compute.manager [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.566963] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.568009] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfb3dc6-f4e1-48a0-ba1e-2613f1a05151 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.575501] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1035.576309] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c359df8-ce8f-49f1-ab40-77fff6b2e3cb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.581407] env[63197]: DEBUG oslo_vmware.api [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1035.581407] env[63197]: value = "task-1364557" [ 1035.581407] env[63197]: _type = "Task" [ 1035.581407] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.589486] env[63197]: DEBUG oslo_vmware.api [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364557, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.635822] env[63197]: DEBUG nova.compute.manager [req-039d881d-e1b1-4fdd-8b91-832de3dccbfe req-473b380e-a8a2-48fb-af2a-ca854773f499 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Received event network-vif-plugged-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.636106] env[63197]: DEBUG oslo_concurrency.lockutils [req-039d881d-e1b1-4fdd-8b91-832de3dccbfe req-473b380e-a8a2-48fb-af2a-ca854773f499 service nova] Acquiring lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.636418] env[63197]: DEBUG oslo_concurrency.lockutils [req-039d881d-e1b1-4fdd-8b91-832de3dccbfe req-473b380e-a8a2-48fb-af2a-ca854773f499 service nova] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.636592] env[63197]: DEBUG oslo_concurrency.lockutils [req-039d881d-e1b1-4fdd-8b91-832de3dccbfe req-473b380e-a8a2-48fb-af2a-ca854773f499 service nova] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.636738] env[63197]: DEBUG nova.compute.manager [req-039d881d-e1b1-4fdd-8b91-832de3dccbfe req-473b380e-a8a2-48fb-af2a-ca854773f499 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] No waiting events found dispatching network-vif-plugged-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.636913] env[63197]: WARNING nova.compute.manager [req-039d881d-e1b1-4fdd-8b91-832de3dccbfe req-473b380e-a8a2-48fb-af2a-ca854773f499 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Received unexpected event network-vif-plugged-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a for instance with vm_state building and task_state spawning. [ 1035.719712] env[63197]: DEBUG nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1035.787964] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "af31d3b0-52ed-408c-8110-c4241e204424" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.788284] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.816674] env[63197]: DEBUG nova.network.neutron [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Successfully updated port: 8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1035.911864] env[63197]: DEBUG nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1036.021658] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364556, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.060037] env[63197]: INFO nova.network.neutron [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Port 44d19552-d824-4f52-93fa-832e41750693 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1036.060439] env[63197]: DEBUG nova.network.neutron [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [{"id": "6de5e286-38ff-452d-b567-9b6c956831d1", "address": "fa:16:3e:31:7e:f3", "network": {"id": "7ac08e03-1288-4a87-9a9c-257ba778c6d6", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-977775153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8db3cc951174f6192ff954ff4d704de", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "298bb8ef-4765-494c-b157-7a349218bd1e", "external-id": "nsx-vlan-transportzone-905", "segmentation_id": 905, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6de5e286-38", "ovs_interfaceid": "6de5e286-38ff-452d-b567-9b6c956831d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.090867] env[63197]: DEBUG oslo_vmware.api [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364557, 'name': PowerOffVM_Task, 'duration_secs': 0.211508} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.091183] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1036.091362] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.091642] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2e7f4bd-fc47-4e90-9b35-264612d555b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.226815] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.227099] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.227224] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleting the datastore file [datastore1] 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.229435] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04e4ec08-cdb0-4f8d-b626-bfdcbdc26709 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.236629] env[63197]: DEBUG oslo_vmware.api [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1036.236629] env[63197]: value = "task-1364559" [ 1036.236629] env[63197]: _type = "Task" [ 1036.236629] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.244141] env[63197]: DEBUG oslo_vmware.api [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364559, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.246257] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.246558] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.248216] env[63197]: INFO nova.compute.claims [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.291749] env[63197]: DEBUG nova.compute.utils [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.316960] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.317169] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.317326] env[63197]: DEBUG nova.network.neutron [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.429828] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.522617] env[63197]: DEBUG oslo_vmware.api [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364556, 'name': RemoveSnapshot_Task, 'duration_secs': 0.708095} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.522975] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1036.523287] env[63197]: INFO nova.compute.manager [None req-dbb93bb6-a993-4dd3-8f0a-ed84c96ed32a tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Took 13.29 seconds to snapshot the instance on the hypervisor. [ 1036.562985] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Releasing lock "refresh_cache-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.746390] env[63197]: DEBUG oslo_vmware.api [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364559, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145241} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.746657] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.746845] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.747059] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.747278] env[63197]: INFO nova.compute.manager [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1036.747521] env[63197]: DEBUG oslo.service.loopingcall [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.747723] env[63197]: DEBUG nova.compute.manager [-] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.747819] env[63197]: DEBUG nova.network.neutron [-] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1036.795056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.865605] env[63197]: DEBUG nova.network.neutron [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1037.066580] env[63197]: DEBUG oslo_concurrency.lockutils [None req-fcfb54bb-207f-4aa6-8e3f-b53b389325ae tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "interface-1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f-44d19552-d824-4f52-93fa-832e41750693" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.014s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.150572] env[63197]: DEBUG nova.network.neutron [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updating instance_info_cache with network_info: [{"id": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "address": "fa:16:3e:cc:28:20", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c51d390-1b", "ovs_interfaceid": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.444941] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ae2b3e-75a0-4999-b6a6-463a058cb014 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.453873] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ead18b-95cf-41d2-a0a1-602fa7bea39a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.487102] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4c39b7-8610-4822-9cba-2379bc580a21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.494562] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a5e8e1-8ef8-4f00-acca-1a086ee8c6e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.511943] env[63197]: DEBUG nova.compute.provider_tree [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.656022] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.656022] env[63197]: DEBUG nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Instance network_info: |[{"id": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "address": "fa:16:3e:cc:28:20", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c51d390-1b", "ovs_interfaceid": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1037.656022] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:28:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c51d390-1bc2-4ffe-90cf-1e43c1593d9a', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.663907] env[63197]: DEBUG oslo.service.loopingcall [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.664496] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1037.664994] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c9d421d-472c-493b-9c85-8b42c3cc23f5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.684757] env[63197]: DEBUG nova.compute.manager [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Received event network-changed-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.685878] env[63197]: DEBUG nova.compute.manager [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Refreshing instance network info cache due to event network-changed-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.685878] env[63197]: DEBUG oslo_concurrency.lockutils [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] Acquiring lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.685878] env[63197]: DEBUG oslo_concurrency.lockutils [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] Acquired lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.685878] env[63197]: DEBUG nova.network.neutron [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Refreshing network info cache for port 8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1037.697255] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.697255] env[63197]: value = "task-1364560" [ 1037.697255] env[63197]: _type = "Task" [ 1037.697255] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.707532] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364560, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.885380] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "af31d3b0-52ed-408c-8110-c4241e204424" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.885529] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.885674] env[63197]: INFO nova.compute.manager [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Attaching volume 7ff408cb-8fad-49a1-8b17-2dd6747d6d12 to /dev/sdb [ 1037.926057] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f7c72d-e0f9-42bf-b573-7c4edd2769b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.934161] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24659fb-bfea-4244-a876-e4ff31538909 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.948284] env[63197]: DEBUG nova.virt.block_device [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Updating existing volume attachment record: c15a3de3-238b-4198-8b02-21c3313b48ec {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1038.017128] env[63197]: DEBUG nova.scheduler.client.report [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.151011] env[63197]: DEBUG nova.network.neutron [-] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.209038] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364560, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.412540] env[63197]: DEBUG nova.network.neutron [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updated VIF entry in instance network info cache for port 8c51d390-1bc2-4ffe-90cf-1e43c1593d9a. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1038.413051] env[63197]: DEBUG nova.network.neutron [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updating instance_info_cache with network_info: [{"id": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "address": "fa:16:3e:cc:28:20", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c51d390-1b", "ovs_interfaceid": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.522870] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.523709] env[63197]: DEBUG nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1038.529304] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.100s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.530789] env[63197]: INFO nova.compute.claims [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.658132] env[63197]: INFO nova.compute.manager [-] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Took 1.91 seconds to deallocate network for instance. [ 1038.708627] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364560, 'name': CreateVM_Task, 'duration_secs': 0.525021} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.708831] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1038.709465] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.709635] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.709965] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1038.710238] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfa00df1-0b6b-439b-bd12-03cdbe535147 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.715230] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1038.715230] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fb4dd3-85c5-c8db-0b13-6e005bb947e7" [ 1038.715230] env[63197]: _type = "Task" [ 1038.715230] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.728258] env[63197]: DEBUG nova.compute.manager [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1038.728704] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fb4dd3-85c5-c8db-0b13-6e005bb947e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.729818] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-354ebd28-b7a8-4246-b4ad-c48ebf4b3ddd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.916274] env[63197]: DEBUG oslo_concurrency.lockutils [req-ab9adb23-5558-439c-8b1d-cfa2bc19b297 req-927a86eb-67c2-47cd-9eae-14768f00b623 service nova] Releasing lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.035206] env[63197]: DEBUG nova.compute.utils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.038453] env[63197]: DEBUG nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Not allocating networking since 'none' was specified. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1039.166009] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.225648] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52fb4dd3-85c5-c8db-0b13-6e005bb947e7, 'name': SearchDatastore_Task, 'duration_secs': 0.012517} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.225955] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.226212] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1039.226448] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.226596] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.226775] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1039.227051] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36e5fb70-6668-4cab-ba14-88aff6dddc87 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.235478] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1039.235668] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1039.236375] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0034e4dc-d556-4847-9c5e-344c5c78c441 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.241238] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1039.241238] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c662c9-59c7-af1d-a8ad-551c46c7cbff" [ 1039.241238] env[63197]: _type = "Task" [ 1039.241238] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.242254] env[63197]: INFO nova.compute.manager [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] instance snapshotting [ 1039.247955] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6915cc7-f8d0-4648-8d52-53636f364117 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.254489] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c662c9-59c7-af1d-a8ad-551c46c7cbff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.267696] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687405d4-fade-4c72-a17b-9eea6f751f0d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.539670] env[63197]: DEBUG nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.660055] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e6f9c3-9493-4f42-80d6-90240d9f6d7c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.666674] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3215eeb3-3c1a-4b4d-9a2f-827f53aad7b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.699680] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6413c5ed-fc64-4d41-8695-b09b0ffbe1d8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.703214] env[63197]: DEBUG nova.compute.manager [req-0c6affbd-c78b-41c7-b81d-04c983e8823e req-d6f07630-baae-43de-b771-56376b04dfed service nova] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Received event network-vif-deleted-6de5e286-38ff-452d-b567-9b6c956831d1 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.710192] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6de16cc-04ba-473c-8a4b-b71e4a487bb0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.724115] env[63197]: DEBUG nova.compute.provider_tree [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.752083] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c662c9-59c7-af1d-a8ad-551c46c7cbff, 'name': SearchDatastore_Task, 'duration_secs': 0.014231} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.753014] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-795a81af-ae74-454a-b8cb-70676e857cdd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.758501] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1039.758501] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52106390-8fca-5477-f590-049590a7bb29" [ 1039.758501] env[63197]: _type = "Task" [ 1039.758501] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.769682] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52106390-8fca-5477-f590-049590a7bb29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.777485] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1039.777944] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e9a11123-de00-4cfb-a865-5dbcc82f5a93 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.786777] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1039.786777] env[63197]: value = "task-1364564" [ 1039.786777] env[63197]: _type = "Task" [ 1039.786777] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.796930] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364564, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.227016] env[63197]: DEBUG nova.scheduler.client.report [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.269799] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52106390-8fca-5477-f590-049590a7bb29, 'name': SearchDatastore_Task, 'duration_secs': 0.010192} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.270134] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.271028] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 18583a35-9cf4-4ae9-965e-be1fdc90efa2/18583a35-9cf4-4ae9-965e-be1fdc90efa2.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1040.271028] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5308d71-7abd-43d8-b248-c47dcd6b1c99 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.278186] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1040.278186] env[63197]: value = "task-1364565" [ 1040.278186] env[63197]: _type = "Task" [ 1040.278186] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.285690] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364565, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.295496] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364564, 'name': CreateSnapshot_Task, 'duration_secs': 0.471977} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.295756] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1040.296481] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476320c2-b30d-4b32-aa2c-4706154d0b57 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.438668] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.438943] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.552840] env[63197]: DEBUG nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1040.581577] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.581828] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.582034] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.582244] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.582450] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.582623] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.582841] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.583059] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.583306] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.583534] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.583775] env[63197]: DEBUG nova.virt.hardware [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.585227] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8e67ca-362f-4759-931e-e223bb9d8ea6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.593384] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607937ff-7517-47c8-ba69-8c670fdb8cb2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.607881] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.613895] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Creating folder: Project (1fb1e7c891b9465e9eb568350e0f39a5). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1040.614340] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7050e5af-c0aa-48fc-a227-7e3b3082e079 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.626615] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Created folder: Project (1fb1e7c891b9465e9eb568350e0f39a5) in parent group-v290286. [ 1040.626977] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Creating folder: Instances. Parent ref: group-v290405. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1040.627307] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b9e2d63-cea5-461c-961e-053ce4f23c68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.641056] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Created folder: Instances in parent group-v290405. [ 1040.641372] env[63197]: DEBUG oslo.service.loopingcall [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.641584] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1040.641807] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe539b8e-bc7b-4baf-98a5-90efe5374b38 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.661402] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.661402] env[63197]: value = "task-1364568" [ 1040.661402] env[63197]: _type = "Task" [ 1040.661402] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.671088] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364568, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.733083] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.733728] env[63197]: DEBUG nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1040.736471] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.571s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.736708] env[63197]: DEBUG nova.objects.instance [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'resources' on Instance uuid 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.787628] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364565, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.813308] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1040.813558] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-14e42ed6-aced-4b90-b575-e4d1aae53c4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.821800] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1040.821800] env[63197]: value = "task-1364569" [ 1040.821800] env[63197]: _type = "Task" [ 1040.821800] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.829907] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364569, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.943067] env[63197]: INFO nova.compute.manager [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Detaching volume 652f45ea-384f-474b-9fd4-23d569d7d473 [ 1040.983438] env[63197]: INFO nova.virt.block_device [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Attempting to driver detach volume 652f45ea-384f-474b-9fd4-23d569d7d473 from mountpoint /dev/sdb [ 1040.983741] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1040.984022] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290384', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'name': 'volume-652f45ea-384f-474b-9fd4-23d569d7d473', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '180e0da7-f7ee-4fcd-be95-c2bf679278d3', 'attached_at': '', 'detached_at': '', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'serial': '652f45ea-384f-474b-9fd4-23d569d7d473'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1040.984940] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfc81b2-7665-4cca-a5ca-10e156246e01 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.008029] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac26cba-13d1-46d8-9125-3165b9d5507e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.016106] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cbf6411-8e78-440b-ad33-d73b33e6301c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.046504] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd50da3-1d2c-4957-b1a9-f5f140d556ea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.068312] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] The volume has not been displaced from its original location: [datastore1] volume-652f45ea-384f-474b-9fd4-23d569d7d473/volume-652f45ea-384f-474b-9fd4-23d569d7d473.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1041.076608] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1041.077019] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0398dda-b3e3-4c9a-af03-98c38cb6a0bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.102908] env[63197]: DEBUG oslo_vmware.api [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1041.102908] env[63197]: value = "task-1364570" [ 1041.102908] env[63197]: _type = "Task" [ 1041.102908] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.113828] env[63197]: DEBUG oslo_vmware.api [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364570, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.170541] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364568, 'name': CreateVM_Task, 'duration_secs': 0.273419} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.170713] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1041.171193] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.171324] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.171669] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1041.171922] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82c45c3b-3a14-4cc6-91e2-4d2ce3482e88 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.176412] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1041.176412] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aae47e-39a1-dea3-b98c-c327355712b6" [ 1041.176412] env[63197]: _type = "Task" [ 1041.176412] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.183647] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aae47e-39a1-dea3-b98c-c327355712b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.239213] env[63197]: DEBUG nova.compute.utils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1041.243217] env[63197]: DEBUG nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Not allocating networking since 'none' was specified. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1041.293665] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364565, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.332323] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364569, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.362830] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a3c058-ebb1-4d5b-a408-8fa4d81dce3f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.370204] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea466dd8-10c6-45d0-b82b-7321306ddce7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.400571] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb73631-4750-4961-9a8b-c9398f87ba80 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.407568] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc6d99f-1a0b-4b09-b245-a0ddb3c37813 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.420225] env[63197]: DEBUG nova.compute.provider_tree [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.613166] env[63197]: DEBUG oslo_vmware.api [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364570, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.686319] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52aae47e-39a1-dea3-b98c-c327355712b6, 'name': SearchDatastore_Task, 'duration_secs': 0.008982} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.686630] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.686864] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1041.687110] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.687267] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.687451] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1041.687709] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fff78952-8e35-49c7-afde-81377142fea1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.695421] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1041.695584] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1041.696265] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7766ab48-16c5-4ab8-9110-7dc01c419e3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.700739] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1041.700739] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ed839b-e7bf-187e-2c6b-d0cd9752aec3" [ 1041.700739] env[63197]: _type = "Task" [ 1041.700739] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.708160] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ed839b-e7bf-187e-2c6b-d0cd9752aec3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.744907] env[63197]: DEBUG nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1041.791517] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364565, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.473613} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.791785] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 18583a35-9cf4-4ae9-965e-be1fdc90efa2/18583a35-9cf4-4ae9-965e-be1fdc90efa2.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1041.791996] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1041.792301] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-deea3839-115d-434f-8a74-145ab831dd54 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.798499] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1041.798499] env[63197]: value = "task-1364571" [ 1041.798499] env[63197]: _type = "Task" [ 1041.798499] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.805853] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364571, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.830807] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364569, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.923331] env[63197]: DEBUG nova.scheduler.client.report [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.117789] env[63197]: DEBUG oslo_vmware.api [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364570, 'name': ReconfigVM_Task, 'duration_secs': 0.862774} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.118257] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1042.126800] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf8464be-812a-46a1-918f-f7a09ed21a18 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.150996] env[63197]: DEBUG oslo_vmware.api [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1042.150996] env[63197]: value = "task-1364572" [ 1042.150996] env[63197]: _type = "Task" [ 1042.150996] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.163613] env[63197]: DEBUG oslo_vmware.api [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364572, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.211572] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ed839b-e7bf-187e-2c6b-d0cd9752aec3, 'name': SearchDatastore_Task, 'duration_secs': 0.008463} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.212441] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61c6a70f-fcf5-4b27-9de0-48be707b585d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.218471] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1042.218471] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526be878-2dda-28d0-1d3c-fb0f998a6eab" [ 1042.218471] env[63197]: _type = "Task" [ 1042.218471] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.227212] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526be878-2dda-28d0-1d3c-fb0f998a6eab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.308262] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364571, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102798} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.308537] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.309330] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c9a18a-c115-4d77-9767-a88bbc918acd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.334192] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 18583a35-9cf4-4ae9-965e-be1fdc90efa2/18583a35-9cf4-4ae9-965e-be1fdc90efa2.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.339166] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a37f45b-a627-4086-a07c-7097d33f3c42 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.360874] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364569, 'name': CloneVM_Task} progress is 95%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.361291] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1042.361291] env[63197]: value = "task-1364573" [ 1042.361291] env[63197]: _type = "Task" [ 1042.361291] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.368671] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364573, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.429279] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.450474] env[63197]: INFO nova.scheduler.client.report [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleted allocations for instance 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f [ 1042.662621] env[63197]: DEBUG oslo_vmware.api [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364572, 'name': ReconfigVM_Task, 'duration_secs': 0.157295} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.662908] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290384', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'name': 'volume-652f45ea-384f-474b-9fd4-23d569d7d473', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '180e0da7-f7ee-4fcd-be95-c2bf679278d3', 'attached_at': '', 'detached_at': '', 'volume_id': '652f45ea-384f-474b-9fd4-23d569d7d473', 'serial': '652f45ea-384f-474b-9fd4-23d569d7d473'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1042.728506] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]526be878-2dda-28d0-1d3c-fb0f998a6eab, 'name': SearchDatastore_Task, 'duration_secs': 0.011998} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.728794] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.729070] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 1b234498-8fe0-475a-9cad-71d0be79307c/1b234498-8fe0-475a-9cad-71d0be79307c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1042.729328] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8e30f1f-da27-498e-9c93-eefb2f9fc468 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.735044] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1042.735044] env[63197]: value = "task-1364575" [ 1042.735044] env[63197]: _type = "Task" [ 1042.735044] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.742081] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364575, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.755213] env[63197]: DEBUG nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1042.776527] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1042.776764] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1042.776923] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1042.777122] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1042.777281] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1042.777431] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1042.777629] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1042.777790] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1042.777957] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1042.778229] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1042.778323] env[63197]: DEBUG nova.virt.hardware [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1042.779104] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c177044f-b5e6-4a11-b604-d675527b6f8f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.786655] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f92c45c-bb3f-4ec4-a66d-dd80414ae01c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.799632] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.805059] env[63197]: DEBUG oslo.service.loopingcall [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.805308] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.805538] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d68c274-ebba-4e86-9ee2-5063946e4c4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.822942] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.822942] env[63197]: value = "task-1364576" [ 1042.822942] env[63197]: _type = "Task" [ 1042.822942] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.833373] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364576, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.836420] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364569, 'name': CloneVM_Task, 'duration_secs': 1.687086} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.836696] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Created linked-clone VM from snapshot [ 1042.837452] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a04b4a5-2327-480f-bb18-6e772f877057 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.844157] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Uploading image 4e88e550-ae87-442d-a232-bbbf7bc9cd0c {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1042.868301] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1042.868301] env[63197]: value = "vm-290408" [ 1042.868301] env[63197]: _type = "VirtualMachine" [ 1042.868301] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1042.868653] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-61afc4a7-5770-4f36-ac48-c5ab006bae06 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.873092] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364573, 'name': ReconfigVM_Task, 'duration_secs': 0.280227} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.873734] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 18583a35-9cf4-4ae9-965e-be1fdc90efa2/18583a35-9cf4-4ae9-965e-be1fdc90efa2.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.874432] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e92b95e-ff09-41fa-be1d-370b45959546 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.880780] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1042.880780] env[63197]: value = "task-1364578" [ 1042.880780] env[63197]: _type = "Task" [ 1042.880780] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.881797] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lease: (returnval){ [ 1042.881797] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52308c1e-a458-ddab-b735-225823d98eb3" [ 1042.881797] env[63197]: _type = "HttpNfcLease" [ 1042.881797] env[63197]: } obtained for exporting VM: (result){ [ 1042.881797] env[63197]: value = "vm-290408" [ 1042.881797] env[63197]: _type = "VirtualMachine" [ 1042.881797] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1042.882247] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the lease: (returnval){ [ 1042.882247] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52308c1e-a458-ddab-b735-225823d98eb3" [ 1042.882247] env[63197]: _type = "HttpNfcLease" [ 1042.882247] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1042.896494] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364578, 'name': Rename_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.898295] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1042.898295] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52308c1e-a458-ddab-b735-225823d98eb3" [ 1042.898295] env[63197]: _type = "HttpNfcLease" [ 1042.898295] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1042.898594] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1042.898594] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52308c1e-a458-ddab-b735-225823d98eb3" [ 1042.898594] env[63197]: _type = "HttpNfcLease" [ 1042.898594] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1042.899427] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c14f7c-9e80-4dbf-bcf7-74730a297566 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.907191] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55cb9-9b67-2d24-446f-9e170b1ea4e1/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1042.907484] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55cb9-9b67-2d24-446f-9e170b1ea4e1/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1042.971900] env[63197]: DEBUG oslo_concurrency.lockutils [None req-0010829e-9da4-44fc-b52b-de098154fcb7 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.410s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.013059] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5dee262e-f61e-4f84-ab2b-9d3cc1608c22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.215924] env[63197]: DEBUG nova.objects.instance [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'flavor' on Instance uuid 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.246292] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364575, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477174} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.246562] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 1b234498-8fe0-475a-9cad-71d0be79307c/1b234498-8fe0-475a-9cad-71d0be79307c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1043.246772] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1043.247040] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c3245aa-1bd4-4e50-a47c-31729138e236 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.253123] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1043.253123] env[63197]: value = "task-1364579" [ 1043.253123] env[63197]: _type = "Task" [ 1043.253123] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.261830] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364579, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.332780] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364576, 'name': CreateVM_Task, 'duration_secs': 0.303112} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.334334] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1043.334334] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.334334] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.334334] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.334609] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed4e186b-2bc1-4677-be36-deb0cf0cfefd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.338963] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1043.338963] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c8722e-8447-b1c2-52b5-82dbedd0eaa3" [ 1043.338963] env[63197]: _type = "Task" [ 1043.338963] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.346679] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c8722e-8447-b1c2-52b5-82dbedd0eaa3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.390790] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364578, 'name': Rename_Task, 'duration_secs': 0.264754} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.391228] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1043.391573] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2731fec-ecf1-47de-9419-f369821bff50 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.398381] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1043.398381] env[63197]: value = "task-1364580" [ 1043.398381] env[63197]: _type = "Task" [ 1043.398381] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.406753] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364580, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.647887] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "4672f595-e1f6-4400-b5a1-065598584980" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.648455] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "4672f595-e1f6-4400-b5a1-065598584980" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.648823] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "4672f595-e1f6-4400-b5a1-065598584980-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.649128] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "4672f595-e1f6-4400-b5a1-065598584980-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.649450] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "4672f595-e1f6-4400-b5a1-065598584980-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.651753] env[63197]: INFO nova.compute.manager [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Terminating instance [ 1043.654070] env[63197]: DEBUG nova.compute.manager [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1043.654717] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1043.655618] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2a9402-0c68-402d-9230-d0eec7e39133 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.664285] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.664725] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab43991e-5857-4564-992b-600bb4356d48 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.670900] env[63197]: DEBUG oslo_vmware.api [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1043.670900] env[63197]: value = "task-1364581" [ 1043.670900] env[63197]: _type = "Task" [ 1043.670900] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.679559] env[63197]: DEBUG oslo_vmware.api [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364581, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.763813] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364579, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072657} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.764261] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1043.765189] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d73729-1c72-415c-8e60-b6ab53eeb71c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.786586] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 1b234498-8fe0-475a-9cad-71d0be79307c/1b234498-8fe0-475a-9cad-71d0be79307c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.786978] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35e5fdf6-6b77-46e6-b9db-a31873f5cd55 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.807881] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1043.807881] env[63197]: value = "task-1364582" [ 1043.807881] env[63197]: _type = "Task" [ 1043.807881] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.817079] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364582, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.849461] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c8722e-8447-b1c2-52b5-82dbedd0eaa3, 'name': SearchDatastore_Task, 'duration_secs': 0.010518} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.849941] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.850297] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.850677] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.850941] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.851458] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.851861] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7606deb-db96-4c6d-9023-4bafb73341ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.860966] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.861313] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.862310] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8b0a1ae-20eb-45b9-ab72-f57a40c15f12 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.868539] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1043.868539] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5241a0ec-d8f1-738c-7cab-b44f330bb3ec" [ 1043.868539] env[63197]: _type = "Task" [ 1043.868539] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.877019] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5241a0ec-d8f1-738c-7cab-b44f330bb3ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.910305] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364580, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.180717] env[63197]: DEBUG oslo_vmware.api [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364581, 'name': PowerOffVM_Task, 'duration_secs': 0.238932} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.181091] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1044.181326] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1044.183247] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f6d8fef-04c7-4a83-bd83-20b259a03b65 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.226944] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6686a85d-7555-4888-a796-1e4d2c6a1c83 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.787s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.319253] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364582, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.380295] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5241a0ec-d8f1-738c-7cab-b44f330bb3ec, 'name': SearchDatastore_Task, 'duration_secs': 0.009595} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.380295] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-171c8659-537b-4d7e-8d2a-2bae8706f867 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.385486] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1044.385486] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5235ffc1-e121-dd13-b899-7b982ddf03f6" [ 1044.385486] env[63197]: _type = "Task" [ 1044.385486] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.393234] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5235ffc1-e121-dd13-b899-7b982ddf03f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.408659] env[63197]: DEBUG oslo_vmware.api [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364580, 'name': PowerOnVM_Task, 'duration_secs': 0.731638} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.409129] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1044.409363] env[63197]: INFO nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Took 9.29 seconds to spawn the instance on the hypervisor. [ 1044.409677] env[63197]: DEBUG nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.410589] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5819102e-5729-46bd-9b3b-6b20d48e2d1f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.433207] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1044.433534] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1044.433737] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleting the datastore file [datastore2] 4672f595-e1f6-4400-b5a1-065598584980 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.434085] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce542aa0-7014-442b-8b16-073bc4767549 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.440939] env[63197]: DEBUG oslo_vmware.api [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for the task: (returnval){ [ 1044.440939] env[63197]: value = "task-1364584" [ 1044.440939] env[63197]: _type = "Task" [ 1044.440939] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.449950] env[63197]: DEBUG oslo_vmware.api [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364584, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.501453] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1044.501853] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290403', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'name': 'volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31d3b0-52ed-408c-8110-c4241e204424', 'attached_at': '', 'detached_at': '', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'serial': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1044.503030] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0fdfcf-c1ac-4f27-9a16-9ca123386056 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.522662] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5e28f9-1d51-41d3-a121-aa081e82088c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.547813] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12/volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1044.548211] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82ba250d-57f4-4958-9dd8-7c352a620bfe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.566374] env[63197]: DEBUG oslo_vmware.api [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1044.566374] env[63197]: value = "task-1364585" [ 1044.566374] env[63197]: _type = "Task" [ 1044.566374] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.578536] env[63197]: DEBUG oslo_vmware.api [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364585, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.805772] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.806225] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.807250] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.807463] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.807671] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.812993] env[63197]: INFO nova.compute.manager [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Terminating instance [ 1044.815518] env[63197]: DEBUG nova.compute.manager [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1044.815727] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1044.816650] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202c0e50-4c6b-4005-b95a-9050b744c3a2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.822839] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364582, 'name': ReconfigVM_Task, 'duration_secs': 0.543879} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.823663] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 1b234498-8fe0-475a-9cad-71d0be79307c/1b234498-8fe0-475a-9cad-71d0be79307c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1044.824341] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d47d195-d08b-4550-9855-3dad439e409a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.828456] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1044.829088] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-66795b54-4550-4aef-a171-dcc4151c2a7c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.833417] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1044.833417] env[63197]: value = "task-1364586" [ 1044.833417] env[63197]: _type = "Task" [ 1044.833417] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.837923] env[63197]: DEBUG oslo_vmware.api [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1044.837923] env[63197]: value = "task-1364587" [ 1044.837923] env[63197]: _type = "Task" [ 1044.837923] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.844909] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364586, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.850314] env[63197]: DEBUG oslo_vmware.api [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.897404] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5235ffc1-e121-dd13-b899-7b982ddf03f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009407} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.897721] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.898028] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.898383] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-147ef3cd-33f8-4b44-9dde-614c7e2932cd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.904968] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1044.904968] env[63197]: value = "task-1364588" [ 1044.904968] env[63197]: _type = "Task" [ 1044.904968] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.913835] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364588, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.932081] env[63197]: INFO nova.compute.manager [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Took 14.11 seconds to build instance. [ 1044.953309] env[63197]: DEBUG oslo_vmware.api [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Task: {'id': task-1364584, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164528} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.953857] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.954263] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1044.954535] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.954789] env[63197]: INFO nova.compute.manager [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Took 1.30 seconds to destroy the instance on the hypervisor. [ 1044.955169] env[63197]: DEBUG oslo.service.loopingcall [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.955513] env[63197]: DEBUG nova.compute.manager [-] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.955650] env[63197]: DEBUG nova.network.neutron [-] [instance: 4672f595-e1f6-4400-b5a1-065598584980] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1045.077310] env[63197]: DEBUG oslo_vmware.api [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364585, 'name': ReconfigVM_Task, 'duration_secs': 0.406587} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.077709] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12/volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1045.083481] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67df0021-4605-4a4c-8e49-c57114c5148f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.098845] env[63197]: DEBUG oslo_vmware.api [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1045.098845] env[63197]: value = "task-1364589" [ 1045.098845] env[63197]: _type = "Task" [ 1045.098845] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.107762] env[63197]: DEBUG oslo_vmware.api [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364589, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.347299] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364586, 'name': Rename_Task, 'duration_secs': 0.17888} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.347966] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1045.348541] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9636d310-785b-4aef-9e2d-2d8e2474328e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.355486] env[63197]: DEBUG oslo_vmware.api [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364587, 'name': PowerOffVM_Task, 'duration_secs': 0.17466} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.356535] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1045.356781] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1045.357073] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b72ad12-ce62-417e-a26d-0a389b23ef0a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.360752] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1045.360752] env[63197]: value = "task-1364590" [ 1045.360752] env[63197]: _type = "Task" [ 1045.360752] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.369828] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364590, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.415996] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364588, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.434940] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f99452ef-825c-455b-b87f-20c0061f5420 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.620s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.466553] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1045.466903] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1045.467143] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleting the datastore file [datastore2] 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.467550] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58eef2b9-854b-4990-8c85-108d08d45d25 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.475369] env[63197]: DEBUG oslo_vmware.api [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1045.475369] env[63197]: value = "task-1364592" [ 1045.475369] env[63197]: _type = "Task" [ 1045.475369] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.484905] env[63197]: DEBUG oslo_vmware.api [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364592, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.614431] env[63197]: DEBUG oslo_vmware.api [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364589, 'name': ReconfigVM_Task, 'duration_secs': 0.147154} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.614728] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290403', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'name': 'volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31d3b0-52ed-408c-8110-c4241e204424', 'attached_at': '', 'detached_at': '', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'serial': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1045.665276] env[63197]: DEBUG nova.compute.manager [req-271f8dec-9e77-49a2-a621-2cf36442d805 req-b214ed8d-3208-404a-a54f-570a9d1807c6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Received event network-vif-deleted-6085d1e1-acf6-4471-8a56-e050285a562d {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.665487] env[63197]: INFO nova.compute.manager [req-271f8dec-9e77-49a2-a621-2cf36442d805 req-b214ed8d-3208-404a-a54f-570a9d1807c6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Neutron deleted interface 6085d1e1-acf6-4471-8a56-e050285a562d; detaching it from the instance and deleting it from the info cache [ 1045.665657] env[63197]: DEBUG nova.network.neutron [req-271f8dec-9e77-49a2-a621-2cf36442d805 req-b214ed8d-3208-404a-a54f-570a9d1807c6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.871479] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364590, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.917171] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364588, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54449} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.917612] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1045.917949] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.918249] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27763f2c-94b0-4ef5-8999-3357bb64d3aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.924697] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1045.924697] env[63197]: value = "task-1364593" [ 1045.924697] env[63197]: _type = "Task" [ 1045.924697] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.932932] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364593, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.984761] env[63197]: DEBUG oslo_vmware.api [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364592, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146385} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.985031] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1045.985229] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1045.985427] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1045.985615] env[63197]: INFO nova.compute.manager [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1045.985858] env[63197]: DEBUG oslo.service.loopingcall [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.986065] env[63197]: DEBUG nova.compute.manager [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1045.986164] env[63197]: DEBUG nova.network.neutron [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1046.142903] env[63197]: DEBUG nova.network.neutron [-] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.168185] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03ef3d75-65a2-48c3-ac3f-9d90ae24284b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.180704] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0261a399-0382-4f85-bc8f-6e19214d2dd9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.210443] env[63197]: DEBUG nova.compute.manager [req-271f8dec-9e77-49a2-a621-2cf36442d805 req-b214ed8d-3208-404a-a54f-570a9d1807c6 service nova] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Detach interface failed, port_id=6085d1e1-acf6-4471-8a56-e050285a562d, reason: Instance 4672f595-e1f6-4400-b5a1-065598584980 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1046.374275] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364590, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.435271] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364593, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061721} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.435604] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.436392] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89bae7b1-9357-460d-9c3c-c68ad775fb7c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.456363] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.458011] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e51bacdf-79a8-4ef3-9886-6054fcbe5baa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.478011] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1046.478011] env[63197]: value = "task-1364594" [ 1046.478011] env[63197]: _type = "Task" [ 1046.478011] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.488913] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364594, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.648080] env[63197]: INFO nova.compute.manager [-] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Took 1.69 seconds to deallocate network for instance. [ 1046.657416] env[63197]: DEBUG nova.objects.instance [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'flavor' on Instance uuid af31d3b0-52ed-408c-8110-c4241e204424 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.872474] env[63197]: DEBUG oslo_vmware.api [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364590, 'name': PowerOnVM_Task, 'duration_secs': 1.327482} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.872774] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1046.872973] env[63197]: INFO nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Took 6.32 seconds to spawn the instance on the hypervisor. [ 1046.873271] env[63197]: DEBUG nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1046.874106] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7221de3-2fed-4efc-8aa2-feae67b28c77 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.972879] env[63197]: DEBUG nova.network.neutron [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.990168] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364594, 'name': ReconfigVM_Task, 'duration_secs': 0.31716} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.990502] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.991736] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71fbd183-8f32-4a4e-ae18-0da87595225d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.998823] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1046.998823] env[63197]: value = "task-1364595" [ 1046.998823] env[63197]: _type = "Task" [ 1046.998823] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.006632] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364595, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.155467] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.155748] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.156037] env[63197]: DEBUG nova.objects.instance [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lazy-loading 'resources' on Instance uuid 4672f595-e1f6-4400-b5a1-065598584980 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1047.165013] env[63197]: DEBUG oslo_concurrency.lockutils [None req-6cd6aadd-688f-4697-a3c7-ec3d784e6398 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.280s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.351515] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "af31d3b0-52ed-408c-8110-c4241e204424" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.351905] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.352082] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "af31d3b0-52ed-408c-8110-c4241e204424-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.352326] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.352539] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.354824] env[63197]: INFO nova.compute.manager [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Terminating instance [ 1047.356823] env[63197]: DEBUG nova.compute.manager [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.357051] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.357294] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba21f7fc-5126-4015-bd54-e1fd16c7b662 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.365081] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1047.365081] env[63197]: value = "task-1364596" [ 1047.365081] env[63197]: _type = "Task" [ 1047.365081] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.373358] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364596, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.390687] env[63197]: INFO nova.compute.manager [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Took 11.17 seconds to build instance. [ 1047.475659] env[63197]: INFO nova.compute.manager [-] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Took 1.49 seconds to deallocate network for instance. [ 1047.510664] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364595, 'name': Rename_Task, 'duration_secs': 0.141439} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.511082] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.511397] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc31efce-6d07-4047-ada3-83e8da348c05 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.519081] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1047.519081] env[63197]: value = "task-1364597" [ 1047.519081] env[63197]: _type = "Task" [ 1047.519081] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.528171] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.706330] env[63197]: DEBUG nova.compute.manager [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Received event network-changed-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.706483] env[63197]: DEBUG nova.compute.manager [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Refreshing instance network info cache due to event network-changed-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.706691] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] Acquiring lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.706880] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] Acquired lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.707047] env[63197]: DEBUG nova.network.neutron [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Refreshing network info cache for port 8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.783699] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308597d2-96e4-4aaf-a150-eefa5080a8ce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.792205] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35395c86-bce5-4026-8665-678a6b7f8973 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.826287] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1009ab6b-f0d3-4ae5-b1e9-1013491216cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.834638] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb5e5f6-06b6-4bb1-8111-de2932685e71 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.850737] env[63197]: DEBUG nova.compute.provider_tree [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.874628] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364596, 'name': PowerOffVM_Task, 'duration_secs': 0.203781} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.875044] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1047.875137] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1047.875473] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290403', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'name': 'volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31d3b0-52ed-408c-8110-c4241e204424', 'attached_at': '', 'detached_at': '', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'serial': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1047.876102] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8126c7d-9722-4b92-8f0f-366b7ec9da42 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.898399] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d96ce845-90f5-4646-82e2-2833aa2566e5 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "1b234498-8fe0-475a-9cad-71d0be79307c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.681s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.899529] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074a4e3b-ae3a-489f-9385-9763554da070 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.906654] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de81857-3dd5-48ef-bae1-aaf9f8e29cbe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.616291] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.619132] env[63197]: DEBUG nova.scheduler.client.report [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.625751] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc97cd7e-ff5e-479c-98e1-d7dcbf7d3f29 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.641546] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] The volume has not been displaced from its original location: [datastore2] volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12/volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1048.646879] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1048.650449] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be5348ac-28f7-486a-ae57-144d789d7898 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.663826] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.672166] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1048.672166] env[63197]: value = "task-1364598" [ 1048.672166] env[63197]: _type = "Task" [ 1048.672166] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.680499] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364598, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.903070] env[63197]: DEBUG nova.network.neutron [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updated VIF entry in instance network info cache for port 8c51d390-1bc2-4ffe-90cf-1e43c1593d9a. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.903535] env[63197]: DEBUG nova.network.neutron [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updating instance_info_cache with network_info: [{"id": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "address": "fa:16:3e:cc:28:20", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c51d390-1b", "ovs_interfaceid": "8c51d390-1bc2-4ffe-90cf-1e43c1593d9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.127084] env[63197]: DEBUG oslo_vmware.api [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364597, 'name': PowerOnVM_Task, 'duration_secs': 1.403726} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.127387] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1049.127586] env[63197]: INFO nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Took 6.37 seconds to spawn the instance on the hypervisor. [ 1049.127767] env[63197]: DEBUG nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.128566] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b650af-0a77-41e8-b26e-253d6f9201fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.131745] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.976s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.134057] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.518s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.134480] env[63197]: DEBUG nova.objects.instance [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'resources' on Instance uuid 180e0da7-f7ee-4fcd-be95-c2bf679278d3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.159286] env[63197]: INFO nova.scheduler.client.report [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Deleted allocations for instance 4672f595-e1f6-4400-b5a1-065598584980 [ 1049.181437] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364598, 'name': ReconfigVM_Task, 'duration_secs': 0.29275} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.182349] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1049.187174] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af6dfe21-3e89-4671-add6-421af156ce4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.203878] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1049.203878] env[63197]: value = "task-1364599" [ 1049.203878] env[63197]: _type = "Task" [ 1049.203878] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.212829] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364599, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.406571] env[63197]: DEBUG oslo_concurrency.lockutils [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] Releasing lock "refresh_cache-18583a35-9cf4-4ae9-965e-be1fdc90efa2" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.406819] env[63197]: DEBUG nova.compute.manager [req-5d068ee7-3e01-497e-aac7-370234e39f05 req-fa595810-ee7f-43f4-afda-336d21d2db0d service nova] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Received event network-vif-deleted-62e2288c-e61e-467a-bd6d-1ac6fd79cb57 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.651124] env[63197]: INFO nova.compute.manager [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Took 13.23 seconds to build instance. [ 1049.667517] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5698aa89-0249-4af2-9bcc-01aba8876545 tempest-AttachInterfacesTestJSON-172351046 tempest-AttachInterfacesTestJSON-172351046-project-member] Lock "4672f595-e1f6-4400-b5a1-065598584980" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.019s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.717880] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364599, 'name': ReconfigVM_Task, 'duration_secs': 0.154943} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.718317] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290403', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'name': 'volume-7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af31d3b0-52ed-408c-8110-c4241e204424', 'attached_at': '', 'detached_at': '', 'volume_id': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12', 'serial': '7ff408cb-8fad-49a1-8b17-2dd6747d6d12'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1049.718806] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.719393] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c10297d-53cc-4146-b3a3-f70b284e0bcf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.728329] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.728581] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e6dd8a9-6264-409a-8fe4-dfc5f7e392b1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.744717] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64fe3126-28ed-46b3-b523-8eacb27515d6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.751525] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8e7bb8-d7aa-451a-a3a8-082902d91078 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.780281] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ee16c9-ad40-4142-8b59-2ac35838a477 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.787899] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb76466b-72f8-4a31-a340-ce2312ad3312 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.801325] env[63197]: DEBUG nova.compute.provider_tree [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.803666] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.803860] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.804058] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleting the datastore file [datastore2] af31d3b0-52ed-408c-8110-c4241e204424 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.804511] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-754f0f59-93ac-4811-be06-4fae7eceda5a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.810915] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1049.810915] env[63197]: value = "task-1364601" [ 1049.810915] env[63197]: _type = "Task" [ 1049.810915] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.819772] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364601, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.153698] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4558c550-c6bc-48b9-802f-921e03427361 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "2ab00a06-545f-4674-b7e3-37354c7460a4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.745s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.306460] env[63197]: DEBUG nova.scheduler.client.report [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1050.321207] env[63197]: DEBUG oslo_vmware.api [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364601, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162674} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.322059] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.322319] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.322461] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.322609] env[63197]: INFO nova.compute.manager [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Took 2.97 seconds to destroy the instance on the hypervisor. [ 1050.322833] env[63197]: DEBUG oslo.service.loopingcall [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.323030] env[63197]: DEBUG nova.compute.manager [-] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1050.323227] env[63197]: DEBUG nova.network.neutron [-] [instance: af31d3b0-52ed-408c-8110-c4241e204424] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1050.597823] env[63197]: INFO nova.compute.manager [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Rebuilding instance [ 1050.608707] env[63197]: DEBUG nova.compute.manager [req-130ae4f3-78d2-44fe-b9af-199bb87a0bc9 req-514f6e34-4a2b-4102-bd81-522e4cf2c32b service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Received event network-vif-deleted-9b9071c4-4e71-4761-bd7c-ab1367d3f9e6 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.608942] env[63197]: INFO nova.compute.manager [req-130ae4f3-78d2-44fe-b9af-199bb87a0bc9 req-514f6e34-4a2b-4102-bd81-522e4cf2c32b service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Neutron deleted interface 9b9071c4-4e71-4761-bd7c-ab1367d3f9e6; detaching it from the instance and deleting it from the info cache [ 1050.609259] env[63197]: DEBUG nova.network.neutron [req-130ae4f3-78d2-44fe-b9af-199bb87a0bc9 req-514f6e34-4a2b-4102-bd81-522e4cf2c32b service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.642415] env[63197]: DEBUG nova.compute.manager [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.643438] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b471c46a-8a62-47b9-b502-f68b260f7ea9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.812386] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.835632] env[63197]: INFO nova.scheduler.client.report [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted allocations for instance 180e0da7-f7ee-4fcd-be95-c2bf679278d3 [ 1051.079602] env[63197]: DEBUG nova.network.neutron [-] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.112333] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7fb134d-7cac-4e78-b40f-d5555a5290ba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.122169] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0907217c-1ece-445c-a8db-2a84de625423 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.148169] env[63197]: DEBUG nova.compute.manager [req-130ae4f3-78d2-44fe-b9af-199bb87a0bc9 req-514f6e34-4a2b-4102-bd81-522e4cf2c32b service nova] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Detach interface failed, port_id=9b9071c4-4e71-4761-bd7c-ab1367d3f9e6, reason: Instance af31d3b0-52ed-408c-8110-c4241e204424 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1051.153653] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1051.153990] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc54a94c-7c78-4940-8292-6b7657cc05d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.160226] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1051.160226] env[63197]: value = "task-1364602" [ 1051.160226] env[63197]: _type = "Task" [ 1051.160226] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.170135] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364602, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.345555] env[63197]: DEBUG oslo_concurrency.lockutils [None req-c3491fe4-b8af-4033-ba22-7adc5c06253d tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "180e0da7-f7ee-4fcd-be95-c2bf679278d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.539s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.582971] env[63197]: INFO nova.compute.manager [-] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Took 1.26 seconds to deallocate network for instance. [ 1051.671733] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364602, 'name': PowerOffVM_Task, 'duration_secs': 0.193714} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.671970] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1051.672442] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1051.673065] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f0eab2-6047-4559-9d5a-4f76096fa109 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.679859] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1051.680133] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-084e84c6-eca7-46e5-8824-ac7451ff4c36 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.702927] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1051.703190] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1051.703386] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleting the datastore file [datastore1] 2ab00a06-545f-4674-b7e3-37354c7460a4 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1051.703659] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb661627-651c-483c-885a-018c835f4e9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.709778] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1051.709778] env[63197]: value = "task-1364604" [ 1051.709778] env[63197]: _type = "Task" [ 1051.709778] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.717944] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.157945] env[63197]: INFO nova.compute.manager [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Took 0.57 seconds to detach 1 volumes for instance. [ 1052.219990] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109364} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.220324] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.220512] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1052.220690] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1052.668589] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.668875] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.669118] env[63197]: DEBUG nova.objects.instance [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'resources' on Instance uuid af31d3b0-52ed-408c-8110-c4241e204424 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.236830] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.237127] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.252519] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1053.252767] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1053.252927] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1053.253206] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1053.253326] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1053.253470] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1053.253715] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1053.253898] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1053.254082] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1053.254270] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1053.254454] env[63197]: DEBUG nova.virt.hardware [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1053.255518] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35149612-00d5-4817-a935-61d2db725ed5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.266167] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf7b44e-1858-4636-a509-15488f9b9adf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.282020] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Instance VIF info [] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1053.287421] env[63197]: DEBUG oslo.service.loopingcall [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.288443] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1053.289336] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e2bacb-92b9-4835-8c19-6813b97349f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.291708] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf2100db-7e39-4ca2-9213-6a76945f507c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.308143] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab4bd32-46aa-43a1-ba3b-3420a69b1f27 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.311880] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1053.311880] env[63197]: value = "task-1364605" [ 1053.311880] env[63197]: _type = "Task" [ 1053.311880] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.339598] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169dfffb-d638-49b9-9403-f9d9de0e70df {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.345301] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364605, 'name': CreateVM_Task} progress is 15%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.351722] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56637d7-1af5-4a2f-8b52-520b77af7596 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.366980] env[63197]: DEBUG nova.compute.provider_tree [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.739515] env[63197]: DEBUG nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1053.823084] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364605, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.868641] env[63197]: DEBUG nova.scheduler.client.report [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.259832] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.323032] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364605, 'name': CreateVM_Task} progress is 99%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.373267] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.376031] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.116s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.377623] env[63197]: INFO nova.compute.claims [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.394031] env[63197]: INFO nova.scheduler.client.report [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted allocations for instance af31d3b0-52ed-408c-8110-c4241e204424 [ 1054.788701] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55cb9-9b67-2d24-446f-9e170b1ea4e1/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1054.789802] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cbfe56d-2dda-410b-89bd-d9bcc9600f26 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.796200] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55cb9-9b67-2d24-446f-9e170b1ea4e1/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1054.796407] env[63197]: ERROR oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55cb9-9b67-2d24-446f-9e170b1ea4e1/disk-0.vmdk due to incomplete transfer. [ 1054.797058] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-847cdcab-9405-47bf-ad3d-16268132f0c1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.802979] env[63197]: DEBUG oslo_vmware.rw_handles [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b55cb9-9b67-2d24-446f-9e170b1ea4e1/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1054.803215] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Uploaded image 4e88e550-ae87-442d-a232-bbbf7bc9cd0c to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1054.805670] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1054.805904] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b5400d2a-8dd6-48fa-885f-34bd78c65089 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.811439] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1054.811439] env[63197]: value = "task-1364606" [ 1054.811439] env[63197]: _type = "Task" [ 1054.811439] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.821470] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364606, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.825113] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364605, 'name': CreateVM_Task, 'duration_secs': 1.295944} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.825663] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1054.825663] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.825831] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.826475] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1054.826475] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62492a07-bc56-41c3-ba5e-521e9392da16 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.830692] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1054.830692] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f56be4-8839-59c1-559d-390605f462f1" [ 1054.830692] env[63197]: _type = "Task" [ 1054.830692] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.838419] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f56be4-8839-59c1-559d-390605f462f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.900987] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24444d9d-d8fe-484d-af73-a215084e30e2 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "af31d3b0-52ed-408c-8110-c4241e204424" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.549s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.321220] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364606, 'name': Destroy_Task, 'duration_secs': 0.3132} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.321514] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Destroyed the VM [ 1055.321740] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1055.321993] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a7a3e7aa-d856-4918-a374-28dae16bfea8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.328425] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1055.328425] env[63197]: value = "task-1364607" [ 1055.328425] env[63197]: _type = "Task" [ 1055.328425] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.338291] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364607, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.341650] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f56be4-8839-59c1-559d-390605f462f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010554} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.341926] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.342172] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1055.342407] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.342558] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.342736] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1055.342982] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d103d00d-18e7-45bf-998c-a9801b40a920 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.350257] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1055.350434] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1055.351125] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9fee2b9-1756-4c86-8a81-8cb3ba25b85c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.355914] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1055.355914] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f15cc5-2796-428b-17cf-231ed683d614" [ 1055.355914] env[63197]: _type = "Task" [ 1055.355914] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.362951] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f15cc5-2796-428b-17cf-231ed683d614, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.480927] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee725b9-62e3-404f-af94-a5e9d4dcc60f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.489309] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f360fe33-25b0-45f0-9726-6f1ee581dcdc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.523099] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6488cb0e-2358-4a2d-92d9-3a5e6458e795 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.531027] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9542d8e-a491-40f0-aa11-78f22278d570 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.544728] env[63197]: DEBUG nova.compute.provider_tree [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.840737] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364607, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.865877] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f15cc5-2796-428b-17cf-231ed683d614, 'name': SearchDatastore_Task, 'duration_secs': 0.007732} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.866822] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f92f9e13-33b2-48a0-ac9b-6d7f09cfbeab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.871554] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1055.871554] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523e0cf7-1751-7320-00f6-c08cde285fba" [ 1055.871554] env[63197]: _type = "Task" [ 1055.871554] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.878868] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523e0cf7-1751-7320-00f6-c08cde285fba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.047054] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "0a283951-36a0-49b4-9127-fc14fe11fda5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.047054] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.047956] env[63197]: DEBUG nova.scheduler.client.report [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.339354] env[63197]: DEBUG oslo_vmware.api [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364607, 'name': RemoveSnapshot_Task, 'duration_secs': 0.730699} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.339728] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1056.339848] env[63197]: INFO nova.compute.manager [None req-08fb0db5-302c-49cc-ad91-b09f2a2becbc tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Took 17.09 seconds to snapshot the instance on the hypervisor. [ 1056.381494] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]523e0cf7-1751-7320-00f6-c08cde285fba, 'name': SearchDatastore_Task, 'duration_secs': 0.009124} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.381772] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.382040] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1056.382326] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e9946be-ffcb-428d-b8f5-3312b149e3e0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.388704] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1056.388704] env[63197]: value = "task-1364608" [ 1056.388704] env[63197]: _type = "Task" [ 1056.388704] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.396603] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364608, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.551562] env[63197]: DEBUG nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.555789] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.180s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.556634] env[63197]: DEBUG nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1056.898053] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364608, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456632} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.898323] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1056.898559] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.898810] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd19f25b-d981-48b2-a09d-e610428c4525 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.904632] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1056.904632] env[63197]: value = "task-1364609" [ 1056.904632] env[63197]: _type = "Task" [ 1056.904632] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.911476] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364609, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.064333] env[63197]: DEBUG nova.compute.utils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.068022] env[63197]: DEBUG nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1057.068022] env[63197]: DEBUG nova.network.neutron [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1057.083659] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.083921] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.085896] env[63197]: INFO nova.compute.claims [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.118090] env[63197]: DEBUG nova.policy [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0788b47942724901a9af191269c400e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f35fa478269e4de68ae997e79237b25f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1057.364495] env[63197]: DEBUG nova.network.neutron [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Successfully created port: ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1057.414930] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364609, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063029} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.415195] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1057.415984] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61fbdde4-54c5-4a33-a26f-824e40cfdc66 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.435534] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.435811] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-798d817b-f1a1-47cb-8ef4-a4b488499185 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.455186] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1057.455186] env[63197]: value = "task-1364610" [ 1057.455186] env[63197]: _type = "Task" [ 1057.455186] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.462599] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.571535] env[63197]: DEBUG nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1057.967814] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364610, 'name': ReconfigVM_Task, 'duration_secs': 0.288993} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.967814] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 2ab00a06-545f-4674-b7e3-37354c7460a4/2ab00a06-545f-4674-b7e3-37354c7460a4.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.967814] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e45a59fd-b490-4561-89d1-257f6c8d4b1f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.975937] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1057.975937] env[63197]: value = "task-1364611" [ 1057.975937] env[63197]: _type = "Task" [ 1057.975937] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.983981] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364611, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.126522] env[63197]: DEBUG nova.compute.manager [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1058.127597] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31f7e93-1bc6-439f-8d8b-77def247e363 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.196049] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d0bdab-3be2-4984-a09b-e033334fa794 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.201999] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f618ec-c35e-4dd4-9e9d-0ab7dacce4c2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.230273] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8d0b4d-9d0e-45dc-a993-e1b7670cce09 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.237511] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579b598b-6181-4178-a097-6201d071f973 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.250742] env[63197]: DEBUG nova.compute.provider_tree [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.485752] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364611, 'name': Rename_Task, 'duration_secs': 0.127428} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.486164] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1058.486306] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6aa2711b-e92f-423c-be3c-259897ab22db {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.491958] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1058.491958] env[63197]: value = "task-1364612" [ 1058.491958] env[63197]: _type = "Task" [ 1058.491958] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.500291] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364612, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.582749] env[63197]: DEBUG nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1058.604974] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.605245] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.605465] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.605737] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.605944] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.606150] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.606427] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.606654] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.606878] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.607086] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.607333] env[63197]: DEBUG nova.virt.hardware [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.608396] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b1d015-79e3-4de2-9731-e7d49cf62e22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.616841] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025ed906-2f2f-4b4b-a1a4-f32961b24159 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.641145] env[63197]: INFO nova.compute.manager [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] instance snapshotting [ 1058.643982] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be9cdbf-cdfe-428c-945c-194ec9acd533 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.663344] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c198009-0a0e-4d8c-a80f-fdfc330c9c15 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.753502] env[63197]: DEBUG nova.scheduler.client.report [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.773762] env[63197]: DEBUG nova.compute.manager [req-5f212e37-ac35-481e-baa2-9426fd6aa953 req-47915286-d530-4181-8fd2-d79ac3be83c0 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-vif-plugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.774109] env[63197]: DEBUG oslo_concurrency.lockutils [req-5f212e37-ac35-481e-baa2-9426fd6aa953 req-47915286-d530-4181-8fd2-d79ac3be83c0 service nova] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.774453] env[63197]: DEBUG oslo_concurrency.lockutils [req-5f212e37-ac35-481e-baa2-9426fd6aa953 req-47915286-d530-4181-8fd2-d79ac3be83c0 service nova] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.774582] env[63197]: DEBUG oslo_concurrency.lockutils [req-5f212e37-ac35-481e-baa2-9426fd6aa953 req-47915286-d530-4181-8fd2-d79ac3be83c0 service nova] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.774762] env[63197]: DEBUG nova.compute.manager [req-5f212e37-ac35-481e-baa2-9426fd6aa953 req-47915286-d530-4181-8fd2-d79ac3be83c0 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] No waiting events found dispatching network-vif-plugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1058.774938] env[63197]: WARNING nova.compute.manager [req-5f212e37-ac35-481e-baa2-9426fd6aa953 req-47915286-d530-4181-8fd2-d79ac3be83c0 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received unexpected event network-vif-plugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af for instance with vm_state building and task_state spawning. [ 1058.859077] env[63197]: DEBUG nova.network.neutron [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Successfully updated port: ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.001812] env[63197]: DEBUG oslo_vmware.api [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364612, 'name': PowerOnVM_Task, 'duration_secs': 0.43758} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.002075] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1059.002251] env[63197]: DEBUG nova.compute.manager [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1059.002998] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca360fd-8dba-4864-8bec-3cab33962bfb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.174180] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1059.174530] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6ad4da05-2b53-4547-ab80-72ee088dc23c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.181928] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1059.181928] env[63197]: value = "task-1364613" [ 1059.181928] env[63197]: _type = "Task" [ 1059.181928] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.189483] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364613, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.259642] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.175s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.260242] env[63197]: DEBUG nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1059.364317] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.364666] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.364894] env[63197]: DEBUG nova.network.neutron [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.518173] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.518548] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.518842] env[63197]: DEBUG nova.objects.instance [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63197) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1059.692493] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364613, 'name': CreateSnapshot_Task, 'duration_secs': 0.442813} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.692690] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1059.693439] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb47b344-bfec-47a3-90ad-27f1b5a16f34 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.765769] env[63197]: DEBUG nova.compute.utils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.767356] env[63197]: DEBUG nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1059.767532] env[63197]: DEBUG nova.network.neutron [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1059.805265] env[63197]: DEBUG nova.policy [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a8e87b64fe46d7ab7e570d57611119', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ceb1016d6d34bff8880dca42d495377', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.914128] env[63197]: DEBUG nova.network.neutron [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1060.127017] env[63197]: DEBUG nova.network.neutron [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.162916] env[63197]: DEBUG nova.network.neutron [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Successfully created port: 6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1060.214262] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1060.214927] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-01379101-4334-485a-93c8-7e9b71c23d41 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.227153] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1060.227153] env[63197]: value = "task-1364614" [ 1060.227153] env[63197]: _type = "Task" [ 1060.227153] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.237378] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364614, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.270241] env[63197]: DEBUG nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1060.512841] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "2ab00a06-545f-4674-b7e3-37354c7460a4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.513162] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "2ab00a06-545f-4674-b7e3-37354c7460a4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.513418] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "2ab00a06-545f-4674-b7e3-37354c7460a4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.513617] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "2ab00a06-545f-4674-b7e3-37354c7460a4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.513792] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "2ab00a06-545f-4674-b7e3-37354c7460a4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.516071] env[63197]: INFO nova.compute.manager [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Terminating instance [ 1060.517687] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "refresh_cache-2ab00a06-545f-4674-b7e3-37354c7460a4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.517847] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "refresh_cache-2ab00a06-545f-4674-b7e3-37354c7460a4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.518027] env[63197]: DEBUG nova.network.neutron [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1060.528104] env[63197]: DEBUG oslo_concurrency.lockutils [None req-788f4111-ec08-447d-a565-b90c504cff8d tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.629926] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.630429] env[63197]: DEBUG nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Instance network_info: |[{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1060.631092] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:03:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec9b6a1b-f520-4f03-b0ce-51ee8346b1af', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.640617] env[63197]: DEBUG oslo.service.loopingcall [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.640868] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1060.641150] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-059ec7d8-cb90-45cd-bae1-35895adc7b9c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.663051] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.663051] env[63197]: value = "task-1364615" [ 1060.663051] env[63197]: _type = "Task" [ 1060.663051] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.671623] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364615, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.737340] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364614, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.800892] env[63197]: DEBUG nova.compute.manager [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.801189] env[63197]: DEBUG nova.compute.manager [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing instance network info cache due to event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.801506] env[63197]: DEBUG oslo_concurrency.lockutils [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.801721] env[63197]: DEBUG oslo_concurrency.lockutils [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.801975] env[63197]: DEBUG nova.network.neutron [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1061.036895] env[63197]: DEBUG nova.network.neutron [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1061.089644] env[63197]: DEBUG nova.network.neutron [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.173384] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364615, 'name': CreateVM_Task, 'duration_secs': 0.292868} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.173541] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1061.174232] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.174410] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.174742] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1061.175047] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ad732da-d1c5-40ff-bf33-dfa5cff97e11 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.180168] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1061.180168] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a20e1e-4a12-582e-2d45-528710c7e115" [ 1061.180168] env[63197]: _type = "Task" [ 1061.180168] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.187631] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a20e1e-4a12-582e-2d45-528710c7e115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.237826] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364614, 'name': CloneVM_Task, 'duration_secs': 0.956891} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.238121] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Created linked-clone VM from snapshot [ 1061.238850] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02212a1e-5f75-4a31-90b8-b063f7de195f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.245512] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Uploading image d0db8687-032d-4d72-afd3-6b39c4a131fb {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1061.265851] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1061.265851] env[63197]: value = "vm-290412" [ 1061.265851] env[63197]: _type = "VirtualMachine" [ 1061.265851] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1061.266132] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fa5c7c5d-83c9-4165-ab0d-987fd972c122 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.272912] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lease: (returnval){ [ 1061.272912] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528e13cb-6b60-9edc-16a4-44937f2d5c48" [ 1061.272912] env[63197]: _type = "HttpNfcLease" [ 1061.272912] env[63197]: } obtained for exporting VM: (result){ [ 1061.272912] env[63197]: value = "vm-290412" [ 1061.272912] env[63197]: _type = "VirtualMachine" [ 1061.272912] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1061.273218] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the lease: (returnval){ [ 1061.273218] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528e13cb-6b60-9edc-16a4-44937f2d5c48" [ 1061.273218] env[63197]: _type = "HttpNfcLease" [ 1061.273218] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1061.279792] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1061.279792] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528e13cb-6b60-9edc-16a4-44937f2d5c48" [ 1061.279792] env[63197]: _type = "HttpNfcLease" [ 1061.279792] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1061.280882] env[63197]: DEBUG nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1061.300668] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1061.300966] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1061.301214] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1061.301477] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1061.301651] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1061.301822] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1061.302071] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1061.302259] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1061.302450] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1061.302725] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1061.302936] env[63197]: DEBUG nova.virt.hardware [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1061.303856] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c245d999-74be-40f1-b048-09b021b056d0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.314874] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e5412d-4cbd-4e98-bd03-cedaf5e6d89f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.592210] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "refresh_cache-2ab00a06-545f-4674-b7e3-37354c7460a4" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.592746] env[63197]: DEBUG nova.compute.manager [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1061.592994] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1061.593980] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e7b3a9-85eb-42eb-951d-8e2be14bd8f8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.601263] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1061.601514] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8aeb5ca-e5e6-49d3-ab01-17ad654b1c89 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.608228] env[63197]: DEBUG oslo_vmware.api [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1061.608228] env[63197]: value = "task-1364617" [ 1061.608228] env[63197]: _type = "Task" [ 1061.608228] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.616600] env[63197]: DEBUG oslo_vmware.api [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364617, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.625024] env[63197]: DEBUG nova.network.neutron [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updated VIF entry in instance network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1061.625478] env[63197]: DEBUG nova.network.neutron [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.691287] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a20e1e-4a12-582e-2d45-528710c7e115, 'name': SearchDatastore_Task, 'duration_secs': 0.009774} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.691443] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.692440] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.692440] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.692440] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.692440] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.692697] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e13ee1f2-3bb7-4118-b4db-c9cf59140d69 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.700756] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.700939] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1061.702107] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e118af3-10ad-4492-ba7f-a55e565f2f22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.708024] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1061.708024] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529f7c64-5ff1-583b-70de-08336cd2af01" [ 1061.708024] env[63197]: _type = "Task" [ 1061.708024] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.715149] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529f7c64-5ff1-583b-70de-08336cd2af01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.780177] env[63197]: DEBUG nova.network.neutron [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Successfully updated port: 6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.785239] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1061.785239] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528e13cb-6b60-9edc-16a4-44937f2d5c48" [ 1061.785239] env[63197]: _type = "HttpNfcLease" [ 1061.785239] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1061.785568] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1061.785568] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528e13cb-6b60-9edc-16a4-44937f2d5c48" [ 1061.785568] env[63197]: _type = "HttpNfcLease" [ 1061.785568] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1061.786318] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ad5905-0050-4dde-888d-ea53fff697ca {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.794408] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527a1ab4-2fad-8703-a228-d8c0f066027f/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1061.794730] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527a1ab4-2fad-8703-a228-d8c0f066027f/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1061.884572] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f66e535f-8235-4708-9d91-abba50f265ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.118250] env[63197]: DEBUG oslo_vmware.api [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364617, 'name': PowerOffVM_Task, 'duration_secs': 0.119942} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.118559] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1062.118737] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1062.119066] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa4d125d-d929-4757-a722-860da4a1322a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.128389] env[63197]: DEBUG oslo_concurrency.lockutils [req-4f86d38e-3f44-4db5-9f50-0da2f47f3b5d req-2ff9a5d1-bc8d-417b-9125-503b10ef1db1 service nova] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.149320] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1062.149642] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1062.149906] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleting the datastore file [datastore2] 2ab00a06-545f-4674-b7e3-37354c7460a4 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1062.150396] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a139483-9e23-454a-b0ff-e6773ed03efc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.157267] env[63197]: DEBUG oslo_vmware.api [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1062.157267] env[63197]: value = "task-1364619" [ 1062.157267] env[63197]: _type = "Task" [ 1062.157267] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.166226] env[63197]: DEBUG oslo_vmware.api [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364619, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.217980] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529f7c64-5ff1-583b-70de-08336cd2af01, 'name': SearchDatastore_Task, 'duration_secs': 0.008249} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.219140] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ce66b9b-5ea6-4470-a044-edeb97fdb579 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.224434] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1062.224434] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f54227-90a7-300e-4af1-04f1e78a1a6d" [ 1062.224434] env[63197]: _type = "Task" [ 1062.224434] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.232927] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f54227-90a7-300e-4af1-04f1e78a1a6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.283190] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.283697] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.283983] env[63197]: DEBUG nova.network.neutron [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.667559] env[63197]: DEBUG oslo_vmware.api [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364619, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101611} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.667947] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1062.668223] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1062.668439] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1062.668634] env[63197]: INFO nova.compute.manager [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1062.669127] env[63197]: DEBUG oslo.service.loopingcall [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.669502] env[63197]: DEBUG nova.compute.manager [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1062.669681] env[63197]: DEBUG nova.network.neutron [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1062.686581] env[63197]: DEBUG nova.network.neutron [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1062.735256] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52f54227-90a7-300e-4af1-04f1e78a1a6d, 'name': SearchDatastore_Task, 'duration_secs': 0.009694} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.735653] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.735988] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.736318] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd993049-b5a0-43d2-b053-2966b27f88b7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.742903] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1062.742903] env[63197]: value = "task-1364620" [ 1062.742903] env[63197]: _type = "Task" [ 1062.742903] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.750910] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364620, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.824564] env[63197]: DEBUG nova.network.neutron [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1062.830317] env[63197]: DEBUG nova.compute.manager [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Received event network-vif-plugged-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.831258] env[63197]: DEBUG oslo_concurrency.lockutils [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] Acquiring lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.831258] env[63197]: DEBUG oslo_concurrency.lockutils [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.831258] env[63197]: DEBUG oslo_concurrency.lockutils [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.831468] env[63197]: DEBUG nova.compute.manager [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] No waiting events found dispatching network-vif-plugged-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1062.832302] env[63197]: WARNING nova.compute.manager [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Received unexpected event network-vif-plugged-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 for instance with vm_state building and task_state spawning. [ 1062.832302] env[63197]: DEBUG nova.compute.manager [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Received event network-changed-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.832422] env[63197]: DEBUG nova.compute.manager [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Refreshing instance network info cache due to event network-changed-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1062.832596] env[63197]: DEBUG oslo_concurrency.lockutils [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] Acquiring lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.010965] env[63197]: DEBUG nova.network.neutron [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Updating instance_info_cache with network_info: [{"id": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "address": "fa:16:3e:61:c3:6f", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6db9eb84-b2", "ovs_interfaceid": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.189320] env[63197]: DEBUG nova.network.neutron [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.254851] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364620, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491317} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.255228] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.255507] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.255834] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-331c871a-3dd6-480e-affa-d08f38f568f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.262225] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1063.262225] env[63197]: value = "task-1364621" [ 1063.262225] env[63197]: _type = "Task" [ 1063.262225] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.270819] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364621, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.513558] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.514064] env[63197]: DEBUG nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Instance network_info: |[{"id": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "address": "fa:16:3e:61:c3:6f", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6db9eb84-b2", "ovs_interfaceid": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1063.514561] env[63197]: DEBUG oslo_concurrency.lockutils [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] Acquired lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.514841] env[63197]: DEBUG nova.network.neutron [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Refreshing network info cache for port 6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1063.516234] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:c3:6f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.523857] env[63197]: DEBUG oslo.service.loopingcall [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.526984] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.527541] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5214dcd2-d2bc-4f27-a0b5-0a8b53a1de3e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.548129] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.548129] env[63197]: value = "task-1364622" [ 1063.548129] env[63197]: _type = "Task" [ 1063.548129] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.556319] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364622, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.691746] env[63197]: INFO nova.compute.manager [-] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Took 1.02 seconds to deallocate network for instance. [ 1063.749322] env[63197]: DEBUG nova.network.neutron [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Updated VIF entry in instance network info cache for port 6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1063.749698] env[63197]: DEBUG nova.network.neutron [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Updating instance_info_cache with network_info: [{"id": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "address": "fa:16:3e:61:c3:6f", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6db9eb84-b2", "ovs_interfaceid": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.772365] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364621, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060828} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.772640] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.773514] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4559b4-9863-4c2c-b754-6431086171e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.796595] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.797280] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a20be2fa-82a0-4c4c-8e4e-199f5c801887 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.818399] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1063.818399] env[63197]: value = "task-1364623" [ 1063.818399] env[63197]: _type = "Task" [ 1063.818399] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.826276] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.059079] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364622, 'name': CreateVM_Task, 'duration_secs': 0.360264} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.059079] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.059509] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.059730] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.060098] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1064.060394] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d6348b7-fc2a-42a3-84e6-e80c797fd767 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.065214] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1064.065214] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5292eaa5-7474-0091-ecc9-b88e4e5a4a43" [ 1064.065214] env[63197]: _type = "Task" [ 1064.065214] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.073583] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5292eaa5-7474-0091-ecc9-b88e4e5a4a43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.199048] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.199395] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.199660] env[63197]: DEBUG nova.objects.instance [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lazy-loading 'resources' on Instance uuid 2ab00a06-545f-4674-b7e3-37354c7460a4 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.252787] env[63197]: DEBUG oslo_concurrency.lockutils [req-8841b72d-59a1-41c2-8e5a-c012e201d660 req-a3b4eba5-e692-40fe-a0a9-bc62b3eb9fa5 service nova] Releasing lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.329735] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364623, 'name': ReconfigVM_Task, 'duration_secs': 0.329731} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.330260] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Reconfigured VM instance instance-00000063 to attach disk [datastore1] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.330882] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d34849f-f5fb-4e71-9d91-39c2b700153b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.337882] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1064.337882] env[63197]: value = "task-1364624" [ 1064.337882] env[63197]: _type = "Task" [ 1064.337882] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.349008] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364624, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.578039] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5292eaa5-7474-0091-ecc9-b88e4e5a4a43, 'name': SearchDatastore_Task, 'duration_secs': 0.012478} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.578039] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.578039] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.578039] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.578039] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.578039] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.578438] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b2c5fd8-c04d-4b6e-8c42-29b9aeacdbb7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.586546] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.586805] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.587584] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c327e56e-5f3e-43b8-83c0-11cf97ebd3c3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.593063] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1064.593063] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b7de5f-0882-fd09-8ffd-e8d2de765194" [ 1064.593063] env[63197]: _type = "Task" [ 1064.593063] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.603124] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b7de5f-0882-fd09-8ffd-e8d2de765194, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.819020] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e40b658-49de-4815-9ffb-3cacc48a715d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.827224] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9cb220-a020-45ac-94a2-31ca78b77058 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.859723] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ce4c82-6850-4a6e-b426-9fb61c34cca0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.867156] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364624, 'name': Rename_Task, 'duration_secs': 0.140819} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.869180] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.869467] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-130a9e1c-f14b-42e0-9cb8-f279c3fe8f7a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.871783] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfe5bb2-27d9-44e3-a32a-f26dba98a6b6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.884716] env[63197]: DEBUG nova.compute.provider_tree [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.887689] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1064.887689] env[63197]: value = "task-1364625" [ 1064.887689] env[63197]: _type = "Task" [ 1064.887689] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.897726] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364625, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.108768] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52b7de5f-0882-fd09-8ffd-e8d2de765194, 'name': SearchDatastore_Task, 'duration_secs': 0.008916} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.109641] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9ee662e-2713-4a03-a1fb-401a309d12e4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.119095] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1065.119095] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a50ad6-2690-5a54-410e-6f44e2728039" [ 1065.119095] env[63197]: _type = "Task" [ 1065.119095] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.132381] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a50ad6-2690-5a54-410e-6f44e2728039, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.390029] env[63197]: DEBUG nova.scheduler.client.report [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.402171] env[63197]: DEBUG oslo_vmware.api [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364625, 'name': PowerOnVM_Task, 'duration_secs': 0.481108} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.402486] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.402703] env[63197]: INFO nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Took 6.82 seconds to spawn the instance on the hypervisor. [ 1065.402885] env[63197]: DEBUG nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.403677] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8eb0fa2-ce80-43f2-b0b8-a1592ab5b332 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.630108] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a50ad6-2690-5a54-410e-6f44e2728039, 'name': SearchDatastore_Task, 'duration_secs': 0.010598} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.630108] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.630108] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 0a283951-36a0-49b4-9127-fc14fe11fda5/0a283951-36a0-49b4-9127-fc14fe11fda5.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.630430] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72397df9-e6a3-499d-9cbf-67ea20c50ca7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.636872] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1065.636872] env[63197]: value = "task-1364626" [ 1065.636872] env[63197]: _type = "Task" [ 1065.636872] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.644761] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.898051] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.919764] env[63197]: INFO nova.scheduler.client.report [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleted allocations for instance 2ab00a06-545f-4674-b7e3-37354c7460a4 [ 1065.923972] env[63197]: INFO nova.compute.manager [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Took 11.68 seconds to build instance. [ 1066.148167] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364626, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.429985] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8d26e850-0bc0-49ed-82e6-b42dfbccae58 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.193s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.435991] env[63197]: DEBUG oslo_concurrency.lockutils [None req-cae0fdd4-838b-4f64-ac1d-05f406db0dfb tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "2ab00a06-545f-4674-b7e3-37354c7460a4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.923s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.588752] env[63197]: DEBUG nova.compute.manager [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.588968] env[63197]: DEBUG nova.compute.manager [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing instance network info cache due to event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1066.589210] env[63197]: DEBUG oslo_concurrency.lockutils [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.589365] env[63197]: DEBUG oslo_concurrency.lockutils [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.589533] env[63197]: DEBUG nova.network.neutron [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1066.649332] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.760013} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.649605] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 0a283951-36a0-49b4-9127-fc14fe11fda5/0a283951-36a0-49b4-9127-fc14fe11fda5.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1066.649824] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1066.650100] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba6f0537-aecc-4989-ba27-7a7495e13fa7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.656326] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1066.656326] env[63197]: value = "task-1364627" [ 1066.656326] env[63197]: _type = "Task" [ 1066.656326] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.665063] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364627, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.050071] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "1b234498-8fe0-475a-9cad-71d0be79307c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.050071] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "1b234498-8fe0-475a-9cad-71d0be79307c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.050481] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "1b234498-8fe0-475a-9cad-71d0be79307c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.050481] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "1b234498-8fe0-475a-9cad-71d0be79307c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.050565] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "1b234498-8fe0-475a-9cad-71d0be79307c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.052790] env[63197]: INFO nova.compute.manager [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Terminating instance [ 1067.054818] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "refresh_cache-1b234498-8fe0-475a-9cad-71d0be79307c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.055016] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquired lock "refresh_cache-1b234498-8fe0-475a-9cad-71d0be79307c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.055207] env[63197]: DEBUG nova.network.neutron [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1067.166342] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364627, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107346} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.166463] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1067.167212] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40d088a-062b-408d-8d15-c4e917835246 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.188535] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 0a283951-36a0-49b4-9127-fc14fe11fda5/0a283951-36a0-49b4-9127-fc14fe11fda5.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1067.191093] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-905453aa-d190-465b-8638-8c502d26738e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.211480] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1067.211480] env[63197]: value = "task-1364628" [ 1067.211480] env[63197]: _type = "Task" [ 1067.211480] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.219134] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364628, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.578541] env[63197]: DEBUG nova.network.neutron [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1067.584362] env[63197]: DEBUG nova.network.neutron [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updated VIF entry in instance network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.584362] env[63197]: DEBUG nova.network.neutron [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.633523] env[63197]: DEBUG nova.network.neutron [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.722444] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364628, 'name': ReconfigVM_Task, 'duration_secs': 0.266517} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.722444] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 0a283951-36a0-49b4-9127-fc14fe11fda5/0a283951-36a0-49b4-9127-fc14fe11fda5.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.723081] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6fe2efc-1ff7-4b74-87b7-05238a0196c6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.729826] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1067.729826] env[63197]: value = "task-1364629" [ 1067.729826] env[63197]: _type = "Task" [ 1067.729826] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.737739] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364629, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.086613] env[63197]: DEBUG oslo_concurrency.lockutils [req-8d6ac397-d967-4a9b-9e3f-34e6000e9b9c req-8c783a4a-a226-47ea-afc9-76153388c44e service nova] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.136027] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Releasing lock "refresh_cache-1b234498-8fe0-475a-9cad-71d0be79307c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.136506] env[63197]: DEBUG nova.compute.manager [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1068.136715] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.137612] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4085c04c-9eb9-4dd9-8378-93f1602078cb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.145747] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.146017] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdd9cd68-3daf-4f73-aa79-bdb8bbf46a98 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.153039] env[63197]: DEBUG oslo_vmware.api [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1068.153039] env[63197]: value = "task-1364630" [ 1068.153039] env[63197]: _type = "Task" [ 1068.153039] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.161462] env[63197]: DEBUG oslo_vmware.api [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.240073] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364629, 'name': Rename_Task, 'duration_secs': 0.175741} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.240372] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1068.240633] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f16d20e6-454e-49a9-872c-ca6b3351c062 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.246912] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1068.246912] env[63197]: value = "task-1364631" [ 1068.246912] env[63197]: _type = "Task" [ 1068.246912] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.256370] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.663644] env[63197]: DEBUG oslo_vmware.api [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364630, 'name': PowerOffVM_Task, 'duration_secs': 0.122714} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.664040] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.664294] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1068.664610] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-912b0717-52fe-4905-97b4-22a505573545 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.693334] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1068.693674] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1068.693879] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleting the datastore file [datastore2] 1b234498-8fe0-475a-9cad-71d0be79307c {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1068.694338] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07950b10-5e1c-44e3-90fc-78a9d9beb1b8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.701329] env[63197]: DEBUG oslo_vmware.api [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for the task: (returnval){ [ 1068.701329] env[63197]: value = "task-1364633" [ 1068.701329] env[63197]: _type = "Task" [ 1068.701329] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.709518] env[63197]: DEBUG oslo_vmware.api [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364633, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.756882] env[63197]: DEBUG oslo_vmware.api [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364631, 'name': PowerOnVM_Task, 'duration_secs': 0.455729} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.757181] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1068.757390] env[63197]: INFO nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1068.757574] env[63197]: DEBUG nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.758369] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2d9acf-6940-469b-8967-34d553ad9630 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.211105] env[63197]: DEBUG oslo_vmware.api [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Task: {'id': task-1364633, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202628} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.211533] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.211592] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.211750] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.211932] env[63197]: INFO nova.compute.manager [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1069.212210] env[63197]: DEBUG oslo.service.loopingcall [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.212414] env[63197]: DEBUG nova.compute.manager [-] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1069.212512] env[63197]: DEBUG nova.network.neutron [-] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.227545] env[63197]: DEBUG nova.network.neutron [-] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1069.275229] env[63197]: INFO nova.compute.manager [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Took 12.21 seconds to build instance. [ 1069.406659] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527a1ab4-2fad-8703-a228-d8c0f066027f/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1069.407746] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-951906c8-5c65-4487-8a0b-4464ef533df5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.413935] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527a1ab4-2fad-8703-a228-d8c0f066027f/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1069.414118] env[63197]: ERROR oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527a1ab4-2fad-8703-a228-d8c0f066027f/disk-0.vmdk due to incomplete transfer. [ 1069.414337] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-665cf663-f566-432a-ac1b-2e899cc839c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.421134] env[63197]: DEBUG oslo_vmware.rw_handles [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527a1ab4-2fad-8703-a228-d8c0f066027f/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1069.421355] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Uploaded image d0db8687-032d-4d72-afd3-6b39c4a131fb to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1069.423428] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1069.423662] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e8002a99-a1e0-461c-9057-5063e2c65bf8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.428986] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1069.428986] env[63197]: value = "task-1364634" [ 1069.428986] env[63197]: _type = "Task" [ 1069.428986] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.436348] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364634, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.730676] env[63197]: DEBUG nova.network.neutron [-] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.777375] env[63197]: DEBUG oslo_concurrency.lockutils [None req-3680b9e4-59f1-43c6-8e60-1340752da3eb tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.731s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.938884] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364634, 'name': Destroy_Task, 'duration_secs': 0.283757} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.939194] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Destroyed the VM [ 1069.939434] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1069.939697] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-485d27da-e345-4129-913c-fc52fca06652 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.945990] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1069.945990] env[63197]: value = "task-1364635" [ 1069.945990] env[63197]: _type = "Task" [ 1069.945990] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.953907] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364635, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.233686] env[63197]: INFO nova.compute.manager [-] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Took 1.02 seconds to deallocate network for instance. [ 1070.456773] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364635, 'name': RemoveSnapshot_Task} progress is 51%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.741056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.741056] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.741056] env[63197]: DEBUG nova.objects.instance [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lazy-loading 'resources' on Instance uuid 1b234498-8fe0-475a-9cad-71d0be79307c {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.748789] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "0a283951-36a0-49b4-9127-fc14fe11fda5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.749082] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.749268] env[63197]: INFO nova.compute.manager [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Shelving [ 1070.956546] env[63197]: DEBUG oslo_vmware.api [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364635, 'name': RemoveSnapshot_Task, 'duration_secs': 0.747757} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.956841] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1070.957091] env[63197]: INFO nova.compute.manager [None req-802f2949-3ce2-449d-9d87-88f0e73542fa tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Took 12.31 seconds to snapshot the instance on the hypervisor. [ 1071.255802] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1071.256186] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0ddb7a4-a902-4e7d-bc29-4a1942634b81 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.263281] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1071.263281] env[63197]: value = "task-1364636" [ 1071.263281] env[63197]: _type = "Task" [ 1071.263281] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.275063] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.333548] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3812dba-3928-483e-8aa1-2eedbff282ed {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.340910] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e681e3-8042-452b-89e9-f8fa603f3405 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.370347] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ccdd9d2-1d6a-49fb-9738-1bfbe0db58ba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.378815] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467b347b-9f51-4521-a2d6-e7ecec3e8e78 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.392413] env[63197]: DEBUG nova.compute.provider_tree [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.775056] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364636, 'name': PowerOffVM_Task, 'duration_secs': 0.195104} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.775348] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1071.776086] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45cdd70-e623-4541-87ab-41c42e3bde28 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.793727] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4797a9f6-c89c-497a-882e-d55086b81631 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.895405] env[63197]: DEBUG nova.scheduler.client.report [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.303739] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1072.304069] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-47cc1862-31e9-4f60-9a8d-ee571bd97990 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.312113] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1072.312113] env[63197]: value = "task-1364637" [ 1072.312113] env[63197]: _type = "Task" [ 1072.312113] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.319985] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364637, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.400227] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.418415] env[63197]: INFO nova.scheduler.client.report [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Deleted allocations for instance 1b234498-8fe0-475a-9cad-71d0be79307c [ 1072.822429] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364637, 'name': CreateSnapshot_Task, 'duration_secs': 0.372245} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.822676] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1072.823455] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3ee639-7c09-4b34-8555-aff4292ec6dc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.926886] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e0b59815-f0ba-4046-8d89-60327afe3b04 tempest-ServerShowV247Test-98679300 tempest-ServerShowV247Test-98679300-project-member] Lock "1b234498-8fe0-475a-9cad-71d0be79307c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.877s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.341121] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1073.341818] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ed53343a-a1c7-4121-a7ba-9b707f1e3b9b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.350343] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1073.350343] env[63197]: value = "task-1364638" [ 1073.350343] env[63197]: _type = "Task" [ 1073.350343] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.359794] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364638, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.797263] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.797566] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.797795] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.797987] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.798177] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.800478] env[63197]: INFO nova.compute.manager [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Terminating instance [ 1073.802591] env[63197]: DEBUG nova.compute.manager [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1073.802833] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1073.803899] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c51d75-af35-456a-a4f2-25805b2c2207 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.812332] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.812585] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5eff07c-e60f-4dc8-b644-9e91d0d473b3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.818808] env[63197]: DEBUG oslo_vmware.api [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1073.818808] env[63197]: value = "task-1364639" [ 1073.818808] env[63197]: _type = "Task" [ 1073.818808] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.827203] env[63197]: DEBUG oslo_vmware.api [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.862555] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364638, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.328238] env[63197]: DEBUG oslo_vmware.api [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364639, 'name': PowerOffVM_Task, 'duration_secs': 0.2025} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.328524] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1074.328701] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.328953] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e3d9b304-a43b-42d6-b20f-dca0e8f5874d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.361100] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364638, 'name': CloneVM_Task} progress is 95%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.387459] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.387749] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.387964] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Deleting the datastore file [datastore1] a77fe81f-fa44-4639-acbb-bf093b5d4f6c {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.388249] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-405cd832-123c-410a-9482-fb42360a1713 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.395312] env[63197]: DEBUG oslo_vmware.api [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1074.395312] env[63197]: value = "task-1364641" [ 1074.395312] env[63197]: _type = "Task" [ 1074.395312] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.403364] env[63197]: DEBUG oslo_vmware.api [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.862582] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364638, 'name': CloneVM_Task, 'duration_secs': 1.257663} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.864062] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Created linked-clone VM from snapshot [ 1074.864062] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1703400d-4741-4d9b-a635-a03597fcbcc6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.871093] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Uploading image b0791729-84ae-4d0f-afea-ab1c14dcf572 {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1074.894088] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1074.894088] env[63197]: value = "vm-290416" [ 1074.894088] env[63197]: _type = "VirtualMachine" [ 1074.894088] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1074.894375] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-eb068863-fcfc-4e55-8b2d-a08dbaf795a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.904302] env[63197]: DEBUG oslo_vmware.api [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364641, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192076} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.905493] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.905698] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.905886] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.906077] env[63197]: INFO nova.compute.manager [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1074.906320] env[63197]: DEBUG oslo.service.loopingcall [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.906569] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lease: (returnval){ [ 1074.906569] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528ed430-1023-8fe3-5474-1d390b03a06a" [ 1074.906569] env[63197]: _type = "HttpNfcLease" [ 1074.906569] env[63197]: } obtained for exporting VM: (result){ [ 1074.906569] env[63197]: value = "vm-290416" [ 1074.906569] env[63197]: _type = "VirtualMachine" [ 1074.906569] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1074.906806] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the lease: (returnval){ [ 1074.906806] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528ed430-1023-8fe3-5474-1d390b03a06a" [ 1074.906806] env[63197]: _type = "HttpNfcLease" [ 1074.906806] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1074.907032] env[63197]: DEBUG nova.compute.manager [-] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1074.907096] env[63197]: DEBUG nova.network.neutron [-] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1074.914467] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1074.914467] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528ed430-1023-8fe3-5474-1d390b03a06a" [ 1074.914467] env[63197]: _type = "HttpNfcLease" [ 1074.914467] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1075.168614] env[63197]: DEBUG nova.compute.manager [req-e1ddf015-5704-43bc-b23a-28e4f859534c req-15675729-c559-43e5-8e19-de5f62476c35 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Received event network-vif-deleted-9161195e-f24b-453c-b121-a6a965441223 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.168971] env[63197]: INFO nova.compute.manager [req-e1ddf015-5704-43bc-b23a-28e4f859534c req-15675729-c559-43e5-8e19-de5f62476c35 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Neutron deleted interface 9161195e-f24b-453c-b121-a6a965441223; detaching it from the instance and deleting it from the info cache [ 1075.169289] env[63197]: DEBUG nova.network.neutron [req-e1ddf015-5704-43bc-b23a-28e4f859534c req-15675729-c559-43e5-8e19-de5f62476c35 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.417028] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1075.417028] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528ed430-1023-8fe3-5474-1d390b03a06a" [ 1075.417028] env[63197]: _type = "HttpNfcLease" [ 1075.417028] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1075.417566] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1075.417566] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]528ed430-1023-8fe3-5474-1d390b03a06a" [ 1075.417566] env[63197]: _type = "HttpNfcLease" [ 1075.417566] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1075.417974] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0875ac-e3df-45e4-9638-405867dc856b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.425028] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396682-96c7-9696-8224-904df1dab244/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1075.425219] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396682-96c7-9696-8224-904df1dab244/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1075.514341] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4835927b-eebe-4047-86ce-03c5f5770342 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.647693] env[63197]: DEBUG nova.network.neutron [-] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.672095] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8518202-ac87-4155-84ef-d4d2d71ecc67 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.682932] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb941b61-9d9a-4df3-b31e-c027fde581b6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.708932] env[63197]: DEBUG nova.compute.manager [req-e1ddf015-5704-43bc-b23a-28e4f859534c req-15675729-c559-43e5-8e19-de5f62476c35 service nova] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Detach interface failed, port_id=9161195e-f24b-453c-b121-a6a965441223, reason: Instance a77fe81f-fa44-4639-acbb-bf093b5d4f6c could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1076.150600] env[63197]: INFO nova.compute.manager [-] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Took 1.24 seconds to deallocate network for instance. [ 1076.605711] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.605985] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1076.606136] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1076.657507] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.657802] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.658044] env[63197]: DEBUG nova.objects.instance [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lazy-loading 'resources' on Instance uuid a77fe81f-fa44-4639-acbb-bf093b5d4f6c {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.238240] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b551bced-eb0d-4e8e-b081-76c451c3e3e5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.245964] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365bb13d-aee0-4ee0-902b-638648896311 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.276944] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58893af-2b82-4580-b100-97beeacbebeb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.283845] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426595ca-6dfc-4073-b8f5-0012edbc2556 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.296589] env[63197]: DEBUG nova.compute.provider_tree [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.799966] env[63197]: DEBUG nova.scheduler.client.report [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.305392] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.331355] env[63197]: INFO nova.scheduler.client.report [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Deleted allocations for instance a77fe81f-fa44-4639-acbb-bf093b5d4f6c [ 1078.839016] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e894ee68-96fc-43ed-8e49-5add94621dac tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "a77fe81f-fa44-4639-acbb-bf093b5d4f6c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.041s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.121062] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Didn't find any instances for network info cache update. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1079.121062] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.121317] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.121425] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.121598] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.121750] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.121928] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.122096] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1079.122278] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.316585] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "d8919322-1fa7-4fc7-9380-a8b7d283b050" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.316956] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.317275] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "d8919322-1fa7-4fc7-9380-a8b7d283b050-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.317496] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.317704] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.320041] env[63197]: INFO nova.compute.manager [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Terminating instance [ 1079.321903] env[63197]: DEBUG nova.compute.manager [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1079.322116] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1079.323038] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c21be740-e7bc-47b2-9e83-60854e5d558c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.330868] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.331547] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-303f18b8-1e24-4c0c-8530-403b1459258f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.337426] env[63197]: DEBUG oslo_vmware.api [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1079.337426] env[63197]: value = "task-1364643" [ 1079.337426] env[63197]: _type = "Task" [ 1079.337426] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.345112] env[63197]: DEBUG oslo_vmware.api [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364643, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.625580] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.625844] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.626034] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.626232] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1079.627249] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34fbfffd-027d-4c2f-b3e9-c56a14ff1e69 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.636591] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b76fe4-2ebb-42c4-95a6-a2907b553d4c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.650716] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf433029-65f1-45a2-aed4-dde0a09ad666 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.657648] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bdabb6-8639-4b17-92a8-17e89e47c885 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.685920] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180775MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1079.686138] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.686259] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.848486] env[63197]: DEBUG oslo_vmware.api [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364643, 'name': PowerOffVM_Task, 'duration_secs': 0.230675} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.848967] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1079.848967] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1079.849232] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bdf79a6e-5445-4257-b7eb-d56fb0f53b04 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.911533] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1079.911779] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1079.912037] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Deleting the datastore file [datastore1] d8919322-1fa7-4fc7-9380-a8b7d283b050 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1079.912361] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3b04e80-7bdd-4d82-9e3b-3f7ec272cd12 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.918920] env[63197]: DEBUG oslo_vmware.api [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for the task: (returnval){ [ 1079.918920] env[63197]: value = "task-1364645" [ 1079.918920] env[63197]: _type = "Task" [ 1079.918920] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.927719] env[63197]: DEBUG oslo_vmware.api [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364645, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.428912] env[63197]: DEBUG oslo_vmware.api [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Task: {'id': task-1364645, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152816} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.429182] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1080.429374] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1080.429555] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1080.429731] env[63197]: INFO nova.compute.manager [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1080.429976] env[63197]: DEBUG oslo.service.loopingcall [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.430207] env[63197]: DEBUG nova.compute.manager [-] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1080.430290] env[63197]: DEBUG nova.network.neutron [-] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1080.689125] env[63197]: DEBUG nova.compute.manager [req-5e04deb2-df90-4b60-a2e1-53a6aff261a8 req-393c4ae5-1bcc-43f7-a38f-bc4846ba1fd8 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Received event network-vif-deleted-ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.689125] env[63197]: INFO nova.compute.manager [req-5e04deb2-df90-4b60-a2e1-53a6aff261a8 req-393c4ae5-1bcc-43f7-a38f-bc4846ba1fd8 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Neutron deleted interface ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52; detaching it from the instance and deleting it from the info cache [ 1080.689283] env[63197]: DEBUG nova.network.neutron [req-5e04deb2-df90-4b60-a2e1-53a6aff261a8 req-393c4ae5-1bcc-43f7-a38f-bc4846ba1fd8 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.723559] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance d8919322-1fa7-4fc7-9380-a8b7d283b050 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.723765] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 18583a35-9cf4-4ae9-965e-be1fdc90efa2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.723890] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance ad216370-3f6d-4ebe-a956-7dd3bf05511f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.724011] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 0a283951-36a0-49b4-9127-fc14fe11fda5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.724200] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1080.724337] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1080.788458] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c50a22a-2770-427e-adc8-a9be7e8927fe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.797822] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef23901-8292-46c5-9764-0c8333223284 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.828155] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c76b7df-766d-4a19-b2f1-5dec8a902ccd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.836357] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fa4a0a-2ff5-4ed5-952d-2b6eec370d2f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.850044] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.160574] env[63197]: DEBUG nova.network.neutron [-] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.193078] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-38f7d290-c13f-49cf-8c5a-52b2f25ede95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.203238] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e749e63c-e1ba-4904-98f0-8ca796e5ff10 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.230501] env[63197]: DEBUG nova.compute.manager [req-5e04deb2-df90-4b60-a2e1-53a6aff261a8 req-393c4ae5-1bcc-43f7-a38f-bc4846ba1fd8 service nova] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Detach interface failed, port_id=ffe2a69e-8558-4ed5-9ac2-fa5094fb8d52, reason: Instance d8919322-1fa7-4fc7-9380-a8b7d283b050 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1081.353351] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.664577] env[63197]: INFO nova.compute.manager [-] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Took 1.23 seconds to deallocate network for instance. [ 1081.858683] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396682-96c7-9696-8224-904df1dab244/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1081.859444] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1081.859696] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.173s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.860571] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc16bb9-cbb4-485e-af4c-ab1f78d944fe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.866985] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396682-96c7-9696-8224-904df1dab244/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1081.867185] env[63197]: ERROR oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396682-96c7-9696-8224-904df1dab244/disk-0.vmdk due to incomplete transfer. [ 1081.867411] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b169c2fe-df30-4aad-a8d8-a9b6e9a2ada7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.874296] env[63197]: DEBUG oslo_vmware.rw_handles [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52396682-96c7-9696-8224-904df1dab244/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1081.874491] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Uploaded image b0791729-84ae-4d0f-afea-ab1c14dcf572 to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1081.876634] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1081.876865] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-db5cf5e8-8a76-4124-bb7d-f256adecd774 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.882892] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1081.882892] env[63197]: value = "task-1364646" [ 1081.882892] env[63197]: _type = "Task" [ 1081.882892] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.890896] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364646, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.171336] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.171706] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.171936] env[63197]: DEBUG nova.objects.instance [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lazy-loading 'resources' on Instance uuid d8919322-1fa7-4fc7-9380-a8b7d283b050 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.393155] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364646, 'name': Destroy_Task, 'duration_secs': 0.301647} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.393455] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Destroyed the VM [ 1082.393738] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1082.393996] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7493d0e8-5629-402c-958a-ec4ed204a37e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.400212] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1082.400212] env[63197]: value = "task-1364647" [ 1082.400212] env[63197]: _type = "Task" [ 1082.400212] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.407418] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364647, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.746292] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a8fd7c-df41-4021-856f-379913712f27 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.753912] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943d4ec9-8326-41fd-9f1d-dda1564e1b8b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.783697] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aafefee-ef0a-4063-9a66-c43ba30c46aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.790927] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9fbf1a-29d7-4c2b-bc71-694f9d27a6cb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.805020] env[63197]: DEBUG nova.compute.provider_tree [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.910833] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364647, 'name': RemoveSnapshot_Task, 'duration_secs': 0.381629} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.911144] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1082.911458] env[63197]: DEBUG nova.compute.manager [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1082.912247] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db3f4cb-f67a-406b-8d92-27f52058854d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.948544] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.948883] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.308359] env[63197]: DEBUG nova.scheduler.client.report [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.423702] env[63197]: INFO nova.compute.manager [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Shelve offloading [ 1083.425480] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1083.425749] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ae24de69-0b20-49c8-9511-e91826bbaaa0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.433406] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1083.433406] env[63197]: value = "task-1364648" [ 1083.433406] env[63197]: _type = "Task" [ 1083.433406] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.441044] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364648, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.453031] env[63197]: DEBUG nova.compute.utils [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1083.813502] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.642s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.834463] env[63197]: INFO nova.scheduler.client.report [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Deleted allocations for instance d8919322-1fa7-4fc7-9380-a8b7d283b050 [ 1083.944017] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] VM already powered off {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1083.944336] env[63197]: DEBUG nova.compute.manager [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1083.945011] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea8ea03-251c-4c0c-a7d3-069fafb76841 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.950482] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.950654] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.950828] env[63197]: DEBUG nova.network.neutron [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1083.955785] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.342526] env[63197]: DEBUG oslo_concurrency.lockutils [None req-24ae9c87-b793-4242-b557-ed55b8067d7f tempest-ListImageFiltersTestJSON-1291026369 tempest-ListImageFiltersTestJSON-1291026369-project-member] Lock "d8919322-1fa7-4fc7-9380-a8b7d283b050" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.026s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.661082] env[63197]: DEBUG nova.network.neutron [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Updating instance_info_cache with network_info: [{"id": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "address": "fa:16:3e:61:c3:6f", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6db9eb84-b2", "ovs_interfaceid": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.026716] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.027109] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.027392] env[63197]: INFO nova.compute.manager [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Attaching volume 5fde2936-25a4-4640-96d0-4a33b68cbe54 to /dev/sdb [ 1085.062955] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f292f1-8c44-49a7-ba67-241f95a4c843 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.070449] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551d2dc8-27d5-4749-8ccc-42ee83bfae9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.082993] env[63197]: DEBUG nova.virt.block_device [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updating existing volume attachment record: acb51a50-6f68-4b33-a98c-216ce7cbdf8e {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1085.163784] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.477033] env[63197]: DEBUG nova.compute.manager [req-9c966a92-007f-47a6-9370-2f9b30b5014d req-466f91ff-cd62-4f78-a811-5dfacdd86cf4 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Received event network-vif-unplugged-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.477283] env[63197]: DEBUG oslo_concurrency.lockutils [req-9c966a92-007f-47a6-9370-2f9b30b5014d req-466f91ff-cd62-4f78-a811-5dfacdd86cf4 service nova] Acquiring lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.477496] env[63197]: DEBUG oslo_concurrency.lockutils [req-9c966a92-007f-47a6-9370-2f9b30b5014d req-466f91ff-cd62-4f78-a811-5dfacdd86cf4 service nova] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.477678] env[63197]: DEBUG oslo_concurrency.lockutils [req-9c966a92-007f-47a6-9370-2f9b30b5014d req-466f91ff-cd62-4f78-a811-5dfacdd86cf4 service nova] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.477862] env[63197]: DEBUG nova.compute.manager [req-9c966a92-007f-47a6-9370-2f9b30b5014d req-466f91ff-cd62-4f78-a811-5dfacdd86cf4 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] No waiting events found dispatching network-vif-unplugged-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1085.478049] env[63197]: WARNING nova.compute.manager [req-9c966a92-007f-47a6-9370-2f9b30b5014d req-466f91ff-cd62-4f78-a811-5dfacdd86cf4 service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Received unexpected event network-vif-unplugged-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 for instance with vm_state shelved and task_state shelving_offloading. [ 1085.493637] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1085.494963] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412b4bc2-a87e-4e95-afcd-eec5bceac7f0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.503194] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1085.503501] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-daf25e3e-79cb-4349-8fc5-eb745d847574 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.572524] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1085.572765] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1085.573394] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleting the datastore file [datastore1] 0a283951-36a0-49b4-9127-fc14fe11fda5 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.573394] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85604ceb-ff2f-4e5c-bbbe-ddd68979d04b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.579171] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1085.579171] env[63197]: value = "task-1364651" [ 1085.579171] env[63197]: _type = "Task" [ 1085.579171] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.586783] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364651, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.089228] env[63197]: DEBUG oslo_vmware.api [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364651, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159617} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.089501] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.089643] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1086.089822] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1086.125408] env[63197]: INFO nova.scheduler.client.report [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted allocations for instance 0a283951-36a0-49b4-9127-fc14fe11fda5 [ 1086.631153] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.631448] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.631712] env[63197]: DEBUG nova.objects.instance [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'resources' on Instance uuid 0a283951-36a0-49b4-9127-fc14fe11fda5 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.135349] env[63197]: DEBUG nova.objects.instance [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'numa_topology' on Instance uuid 0a283951-36a0-49b4-9127-fc14fe11fda5 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.517387] env[63197]: DEBUG nova.compute.manager [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Received event network-changed-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.517584] env[63197]: DEBUG nova.compute.manager [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Refreshing instance network info cache due to event network-changed-6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1087.517803] env[63197]: DEBUG oslo_concurrency.lockutils [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] Acquiring lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.517949] env[63197]: DEBUG oslo_concurrency.lockutils [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] Acquired lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.518112] env[63197]: DEBUG nova.network.neutron [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Refreshing network info cache for port 6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1087.638867] env[63197]: DEBUG nova.objects.base [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Object Instance<0a283951-36a0-49b4-9127-fc14fe11fda5> lazy-loaded attributes: resources,numa_topology {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1087.687843] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b8600d-c711-428b-bfb1-992e223a7b66 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.696396] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50ebd3d-b9e6-4d36-a3be-2272cce8f777 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.726667] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61705f4c-3e62-4446-a49d-32f09d3d6b9e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.734084] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e386e9-c53e-46ef-ba5f-fe1dd8c322aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.747560] env[63197]: DEBUG nova.compute.provider_tree [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.794230] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "0a283951-36a0-49b4-9127-fc14fe11fda5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.253261] env[63197]: DEBUG nova.scheduler.client.report [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.256556] env[63197]: DEBUG nova.network.neutron [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Updated VIF entry in instance network info cache for port 6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1088.257076] env[63197]: DEBUG nova.network.neutron [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Updating instance_info_cache with network_info: [{"id": "6db9eb84-b2cd-4e9a-a6f1-28ccaf2a3614", "address": "fa:16:3e:61:c3:6f", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": null, "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap6db9eb84-b2", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.759417] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.128s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.762251] env[63197]: DEBUG oslo_concurrency.lockutils [req-72cbab78-eef8-4f7a-a9a8-3e77a17a9314 req-769b5e66-4558-4caa-acb8-93c1ef5b1edd service nova] Releasing lock "refresh_cache-0a283951-36a0-49b4-9127-fc14fe11fda5" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.269761] env[63197]: DEBUG oslo_concurrency.lockutils [None req-58307bf6-96f5-4242-bfab-87c0b13552dd tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 18.521s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.270843] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.477s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.271132] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.271404] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.271617] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.273896] env[63197]: INFO nova.compute.manager [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Terminating instance [ 1089.275731] env[63197]: DEBUG nova.compute.manager [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1089.275968] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1089.276306] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9b1cdd8-0c11-43be-85ee-8a7bdddf8565 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.286855] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3766e9c-fc5c-45bd-b63f-3054ba129c34 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.313269] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0a283951-36a0-49b4-9127-fc14fe11fda5 could not be found. [ 1089.314254] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1089.314254] env[63197]: INFO nova.compute.manager [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1089.314254] env[63197]: DEBUG oslo.service.loopingcall [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.314388] env[63197]: DEBUG nova.compute.manager [-] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1089.314429] env[63197]: DEBUG nova.network.neutron [-] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1089.628088] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1089.628354] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290417', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'name': 'volume-5fde2936-25a4-4640-96d0-4a33b68cbe54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18583a35-9cf4-4ae9-965e-be1fdc90efa2', 'attached_at': '', 'detached_at': '', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'serial': '5fde2936-25a4-4640-96d0-4a33b68cbe54'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1089.629289] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451479a5-bf13-4f0a-a927-3baa09a1afb2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.645417] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acab1f10-486f-434e-a657-a8b6343aba79 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.669141] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] volume-5fde2936-25a4-4640-96d0-4a33b68cbe54/volume-5fde2936-25a4-4640-96d0-4a33b68cbe54.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.669716] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cdce4e3-3d38-4d06-99a3-874d95ef845e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.689895] env[63197]: DEBUG oslo_vmware.api [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1089.689895] env[63197]: value = "task-1364656" [ 1089.689895] env[63197]: _type = "Task" [ 1089.689895] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.698039] env[63197]: DEBUG oslo_vmware.api [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364656, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.027164] env[63197]: DEBUG nova.network.neutron [-] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.199719] env[63197]: DEBUG oslo_vmware.api [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364656, 'name': ReconfigVM_Task, 'duration_secs': 0.343394} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.200013] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Reconfigured VM instance instance-00000060 to attach disk [datastore2] volume-5fde2936-25a4-4640-96d0-4a33b68cbe54/volume-5fde2936-25a4-4640-96d0-4a33b68cbe54.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.204651] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6797b511-63ee-41d2-b510-9ada998e5956 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.219536] env[63197]: DEBUG oslo_vmware.api [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1090.219536] env[63197]: value = "task-1364657" [ 1090.219536] env[63197]: _type = "Task" [ 1090.219536] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.226860] env[63197]: DEBUG oslo_vmware.api [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364657, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.529641] env[63197]: INFO nova.compute.manager [-] [instance: 0a283951-36a0-49b4-9127-fc14fe11fda5] Took 1.21 seconds to deallocate network for instance. [ 1090.729213] env[63197]: DEBUG oslo_vmware.api [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364657, 'name': ReconfigVM_Task, 'duration_secs': 0.127295} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.729570] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290417', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'name': 'volume-5fde2936-25a4-4640-96d0-4a33b68cbe54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18583a35-9cf4-4ae9-965e-be1fdc90efa2', 'attached_at': '', 'detached_at': '', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'serial': '5fde2936-25a4-4640-96d0-4a33b68cbe54'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1091.558610] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd82287f-7ce3-4c3f-a970-99892919fb80 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "0a283951-36a0-49b4-9127-fc14fe11fda5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.288s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.765433] env[63197]: DEBUG nova.objects.instance [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'flavor' on Instance uuid 18583a35-9cf4-4ae9-965e-be1fdc90efa2 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.270839] env[63197]: DEBUG oslo_concurrency.lockutils [None req-d9ad052a-833d-4ff6-ada3-2e9a571df10d tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.243s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.544076] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.544404] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.617235] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.617517] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.047793] env[63197]: INFO nova.compute.manager [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Detaching volume 5fde2936-25a4-4640-96d0-4a33b68cbe54 [ 1093.084307] env[63197]: INFO nova.virt.block_device [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Attempting to driver detach volume 5fde2936-25a4-4640-96d0-4a33b68cbe54 from mountpoint /dev/sdb [ 1093.084562] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1093.084777] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290417', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'name': 'volume-5fde2936-25a4-4640-96d0-4a33b68cbe54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18583a35-9cf4-4ae9-965e-be1fdc90efa2', 'attached_at': '', 'detached_at': '', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'serial': '5fde2936-25a4-4640-96d0-4a33b68cbe54'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1093.085703] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33239581-49e5-44da-b7d1-29a39d631f12 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.106972] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4d0d75-628c-4d0c-b31e-9e93ddd30869 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.113783] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b02411-d3fa-494e-b9f1-a1c35b7349f3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.134018] env[63197]: DEBUG nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1093.137291] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c181cc3-b356-4bcc-8b62-4b3972d432ec {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.151628] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] The volume has not been displaced from its original location: [datastore2] volume-5fde2936-25a4-4640-96d0-4a33b68cbe54/volume-5fde2936-25a4-4640-96d0-4a33b68cbe54.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1093.156905] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1093.157844] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3757140-812a-4376-b5b9-63d1f5a8b5dc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.176027] env[63197]: DEBUG oslo_vmware.api [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1093.176027] env[63197]: value = "task-1364660" [ 1093.176027] env[63197]: _type = "Task" [ 1093.176027] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.184561] env[63197]: DEBUG oslo_vmware.api [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364660, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.658803] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.659098] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.660719] env[63197]: INFO nova.compute.claims [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1093.685476] env[63197]: DEBUG oslo_vmware.api [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364660, 'name': ReconfigVM_Task, 'duration_secs': 0.221749} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.685761] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1093.690594] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-006d3a2d-8256-4534-b974-a2bbf69a3d35 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.705831] env[63197]: DEBUG oslo_vmware.api [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1093.705831] env[63197]: value = "task-1364661" [ 1093.705831] env[63197]: _type = "Task" [ 1093.705831] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.713622] env[63197]: DEBUG oslo_vmware.api [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364661, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.215197] env[63197]: DEBUG oslo_vmware.api [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364661, 'name': ReconfigVM_Task, 'duration_secs': 0.126428} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.215490] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290417', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'name': 'volume-5fde2936-25a4-4640-96d0-4a33b68cbe54', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '18583a35-9cf4-4ae9-965e-be1fdc90efa2', 'attached_at': '', 'detached_at': '', 'volume_id': '5fde2936-25a4-4640-96d0-4a33b68cbe54', 'serial': '5fde2936-25a4-4640-96d0-4a33b68cbe54'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1094.717609] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a31b19f-4844-45da-8629-049ddb1cccff {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.725057] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79317d6a-f34a-4f28-bc15-35b2946cb34c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.755653] env[63197]: DEBUG nova.objects.instance [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'flavor' on Instance uuid 18583a35-9cf4-4ae9-965e-be1fdc90efa2 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.757467] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2883b102-534e-4cd0-a9ca-09b33c4acee5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.764550] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d1b301-3c90-4c23-b047-fec28accd2aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.778045] env[63197]: DEBUG nova.compute.provider_tree [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.280633] env[63197]: DEBUG nova.scheduler.client.report [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.764636] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a8fdbf93-af56-493d-b1c5-24e653466475 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.220s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.784903] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.126s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.785439] env[63197]: DEBUG nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1096.289160] env[63197]: DEBUG nova.compute.utils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1096.290445] env[63197]: DEBUG nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1096.290617] env[63197]: DEBUG nova.network.neutron [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1096.341466] env[63197]: DEBUG nova.policy [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a8e87b64fe46d7ab7e570d57611119', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ceb1016d6d34bff8880dca42d495377', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1096.597843] env[63197]: DEBUG nova.network.neutron [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Successfully created port: c0f5ed33-2425-4178-82ea-7eadf90cc2cf {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1096.778288] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.778560] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.778739] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.778929] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.779121] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.781186] env[63197]: INFO nova.compute.manager [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Terminating instance [ 1096.783884] env[63197]: DEBUG nova.compute.manager [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1096.784105] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1096.784943] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d6caeb-0ba8-4696-b8d2-3da630c97888 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.792549] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1096.793013] env[63197]: DEBUG nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1096.795306] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9fe4afc-9992-4bac-8510-f1b2e1b4f162 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.803188] env[63197]: DEBUG oslo_vmware.api [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1096.803188] env[63197]: value = "task-1364663" [ 1096.803188] env[63197]: _type = "Task" [ 1096.803188] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.810763] env[63197]: DEBUG oslo_vmware.api [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.312955] env[63197]: DEBUG oslo_vmware.api [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364663, 'name': PowerOffVM_Task, 'duration_secs': 0.186883} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.312955] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.312955] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.313069] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ee465434-d667-4903-b099-f7d9e6cafa8b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.373704] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.373930] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.374144] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleting the datastore file [datastore2] 18583a35-9cf4-4ae9-965e-be1fdc90efa2 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.374429] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6da4cb13-5079-44ac-a20a-786ecc62e27e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.381556] env[63197]: DEBUG oslo_vmware.api [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1097.381556] env[63197]: value = "task-1364666" [ 1097.381556] env[63197]: _type = "Task" [ 1097.381556] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.389277] env[63197]: DEBUG oslo_vmware.api [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364666, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.805086] env[63197]: DEBUG nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1097.841599] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1097.841813] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1097.842079] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1097.842459] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1097.843591] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1097.843762] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1097.844063] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1097.844278] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1097.844497] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1097.844697] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1097.844917] env[63197]: DEBUG nova.virt.hardware [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1097.846268] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ccf26d-e29f-4473-aa26-60ff155d8525 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.856658] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62009b73-0932-4dec-af01-64a4263eb849 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.891297] env[63197]: DEBUG oslo_vmware.api [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364666, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146737} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.891568] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1097.891848] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1097.892161] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1097.892374] env[63197]: INFO nova.compute.manager [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1097.892629] env[63197]: DEBUG oslo.service.loopingcall [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.892827] env[63197]: DEBUG nova.compute.manager [-] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1097.892923] env[63197]: DEBUG nova.network.neutron [-] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.006046] env[63197]: DEBUG nova.compute.manager [req-6c67ee6b-5f72-4244-b89b-1195c818add3 req-ce34ae74-de59-4c88-8e0d-8c5217dc6e6b service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Received event network-vif-plugged-c0f5ed33-2425-4178-82ea-7eadf90cc2cf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1098.006505] env[63197]: DEBUG oslo_concurrency.lockutils [req-6c67ee6b-5f72-4244-b89b-1195c818add3 req-ce34ae74-de59-4c88-8e0d-8c5217dc6e6b service nova] Acquiring lock "87571a6e-9633-4d94-9cec-fbb68c53afbb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.006505] env[63197]: DEBUG oslo_concurrency.lockutils [req-6c67ee6b-5f72-4244-b89b-1195c818add3 req-ce34ae74-de59-4c88-8e0d-8c5217dc6e6b service nova] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.006729] env[63197]: DEBUG oslo_concurrency.lockutils [req-6c67ee6b-5f72-4244-b89b-1195c818add3 req-ce34ae74-de59-4c88-8e0d-8c5217dc6e6b service nova] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.006843] env[63197]: DEBUG nova.compute.manager [req-6c67ee6b-5f72-4244-b89b-1195c818add3 req-ce34ae74-de59-4c88-8e0d-8c5217dc6e6b service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] No waiting events found dispatching network-vif-plugged-c0f5ed33-2425-4178-82ea-7eadf90cc2cf {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1098.007031] env[63197]: WARNING nova.compute.manager [req-6c67ee6b-5f72-4244-b89b-1195c818add3 req-ce34ae74-de59-4c88-8e0d-8c5217dc6e6b service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Received unexpected event network-vif-plugged-c0f5ed33-2425-4178-82ea-7eadf90cc2cf for instance with vm_state building and task_state spawning. [ 1098.465658] env[63197]: DEBUG nova.compute.manager [req-1ff8bc07-32dc-4e2c-b264-56e6068bbfe7 req-77fd25c6-c1da-40c5-936c-1b56c2100bec service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Received event network-vif-deleted-8c51d390-1bc2-4ffe-90cf-1e43c1593d9a {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1098.465658] env[63197]: INFO nova.compute.manager [req-1ff8bc07-32dc-4e2c-b264-56e6068bbfe7 req-77fd25c6-c1da-40c5-936c-1b56c2100bec service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Neutron deleted interface 8c51d390-1bc2-4ffe-90cf-1e43c1593d9a; detaching it from the instance and deleting it from the info cache [ 1098.465832] env[63197]: DEBUG nova.network.neutron [req-1ff8bc07-32dc-4e2c-b264-56e6068bbfe7 req-77fd25c6-c1da-40c5-936c-1b56c2100bec service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.622704] env[63197]: DEBUG nova.network.neutron [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Successfully updated port: c0f5ed33-2425-4178-82ea-7eadf90cc2cf {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1098.944972] env[63197]: DEBUG nova.network.neutron [-] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.968350] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a30f88c-e3d9-4e8a-b2b6-f913ffce37f7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.977644] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba92d03-1472-496e-9ccf-3cfe8c9903d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.005489] env[63197]: DEBUG nova.compute.manager [req-1ff8bc07-32dc-4e2c-b264-56e6068bbfe7 req-77fd25c6-c1da-40c5-936c-1b56c2100bec service nova] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Detach interface failed, port_id=8c51d390-1bc2-4ffe-90cf-1e43c1593d9a, reason: Instance 18583a35-9cf4-4ae9-965e-be1fdc90efa2 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1099.125850] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-87571a6e-9633-4d94-9cec-fbb68c53afbb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.126033] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-87571a6e-9633-4d94-9cec-fbb68c53afbb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.126077] env[63197]: DEBUG nova.network.neutron [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1099.448262] env[63197]: INFO nova.compute.manager [-] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Took 1.56 seconds to deallocate network for instance. [ 1099.662024] env[63197]: DEBUG nova.network.neutron [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1099.795695] env[63197]: DEBUG nova.network.neutron [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Updating instance_info_cache with network_info: [{"id": "c0f5ed33-2425-4178-82ea-7eadf90cc2cf", "address": "fa:16:3e:dd:ce:b0", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0f5ed33-24", "ovs_interfaceid": "c0f5ed33-2425-4178-82ea-7eadf90cc2cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.956091] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.956412] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.956646] env[63197]: DEBUG nova.objects.instance [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'resources' on Instance uuid 18583a35-9cf4-4ae9-965e-be1fdc90efa2 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.298968] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-87571a6e-9633-4d94-9cec-fbb68c53afbb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.299459] env[63197]: DEBUG nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Instance network_info: |[{"id": "c0f5ed33-2425-4178-82ea-7eadf90cc2cf", "address": "fa:16:3e:dd:ce:b0", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0f5ed33-24", "ovs_interfaceid": "c0f5ed33-2425-4178-82ea-7eadf90cc2cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1100.300064] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:ce:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0f5ed33-2425-4178-82ea-7eadf90cc2cf', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1100.311875] env[63197]: DEBUG oslo.service.loopingcall [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1100.312115] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1100.312355] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d44eebaf-4684-4710-9771-10a968194c13 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.331130] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1100.331130] env[63197]: value = "task-1364668" [ 1100.331130] env[63197]: _type = "Task" [ 1100.331130] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.338863] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364668, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.493731] env[63197]: DEBUG nova.compute.manager [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Received event network-changed-c0f5ed33-2425-4178-82ea-7eadf90cc2cf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1100.493972] env[63197]: DEBUG nova.compute.manager [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Refreshing instance network info cache due to event network-changed-c0f5ed33-2425-4178-82ea-7eadf90cc2cf. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1100.494216] env[63197]: DEBUG oslo_concurrency.lockutils [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] Acquiring lock "refresh_cache-87571a6e-9633-4d94-9cec-fbb68c53afbb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.494367] env[63197]: DEBUG oslo_concurrency.lockutils [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] Acquired lock "refresh_cache-87571a6e-9633-4d94-9cec-fbb68c53afbb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.494527] env[63197]: DEBUG nova.network.neutron [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Refreshing network info cache for port c0f5ed33-2425-4178-82ea-7eadf90cc2cf {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1100.521899] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7012540-8dfb-4b35-975f-9fc9b4f83a3f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.528997] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4514d07b-19bd-4869-b882-a5caca31f52d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.562843] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0ffcce-7e93-4977-b071-24d5500f9889 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.570338] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15aa45c3-0353-49ba-af8a-1f9774289670 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.583456] env[63197]: DEBUG nova.compute.provider_tree [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.841391] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364668, 'name': CreateVM_Task, 'duration_secs': 0.314507} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.841539] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1100.842213] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.842432] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.842783] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1100.843052] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89891610-bd92-489f-b981-1ef94ad99150 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.847225] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1100.847225] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e16b46-8203-b952-08cf-29150efa945e" [ 1100.847225] env[63197]: _type = "Task" [ 1100.847225] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.854674] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e16b46-8203-b952-08cf-29150efa945e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.087522] env[63197]: DEBUG nova.scheduler.client.report [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.177552] env[63197]: DEBUG nova.network.neutron [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Updated VIF entry in instance network info cache for port c0f5ed33-2425-4178-82ea-7eadf90cc2cf. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1101.177918] env[63197]: DEBUG nova.network.neutron [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Updating instance_info_cache with network_info: [{"id": "c0f5ed33-2425-4178-82ea-7eadf90cc2cf", "address": "fa:16:3e:dd:ce:b0", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0f5ed33-24", "ovs_interfaceid": "c0f5ed33-2425-4178-82ea-7eadf90cc2cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.356933] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52e16b46-8203-b952-08cf-29150efa945e, 'name': SearchDatastore_Task, 'duration_secs': 0.010465} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.357212] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.357448] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1101.357686] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.357836] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.358087] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1101.358361] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8c46f8a6-cccc-4a5c-8cdd-c82c43f26b04 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.366684] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1101.366863] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1101.367555] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e44598a2-fd17-491c-9c12-e61d62534d58 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.372278] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1101.372278] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bc0fc4-3188-3676-e13e-53dc06765930" [ 1101.372278] env[63197]: _type = "Task" [ 1101.372278] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.379155] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bc0fc4-3188-3676-e13e-53dc06765930, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.591508] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.676367] env[63197]: INFO nova.scheduler.client.report [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted allocations for instance 18583a35-9cf4-4ae9-965e-be1fdc90efa2 [ 1101.680374] env[63197]: DEBUG oslo_concurrency.lockutils [req-9a54adb0-4cab-466a-b223-7f8302672f60 req-39f79ac4-6aa9-48aa-914f-d98df8adb5cf service nova] Releasing lock "refresh_cache-87571a6e-9633-4d94-9cec-fbb68c53afbb" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.882361] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52bc0fc4-3188-3676-e13e-53dc06765930, 'name': SearchDatastore_Task, 'duration_secs': 0.008343} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.883143] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1d4c453-d644-47cd-9546-91d0223e6fc0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.888274] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1101.888274] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520780b3-decf-86f7-42a6-1cfd5af7310a" [ 1101.888274] env[63197]: _type = "Task" [ 1101.888274] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.895418] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520780b3-decf-86f7-42a6-1cfd5af7310a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.184451] env[63197]: DEBUG oslo_concurrency.lockutils [None req-e75f6d96-611e-4095-8d10-e52a99a775e3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "18583a35-9cf4-4ae9-965e-be1fdc90efa2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.406s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.399743] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520780b3-decf-86f7-42a6-1cfd5af7310a, 'name': SearchDatastore_Task, 'duration_secs': 0.010992} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.400016] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1102.400285] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 87571a6e-9633-4d94-9cec-fbb68c53afbb/87571a6e-9633-4d94-9cec-fbb68c53afbb.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1102.400552] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1043e75-34ec-4953-9e97-95e098ea5403 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.407260] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1102.407260] env[63197]: value = "task-1364670" [ 1102.407260] env[63197]: _type = "Task" [ 1102.407260] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.414536] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364670, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.916976] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364670, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460229} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.917307] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 87571a6e-9633-4d94-9cec-fbb68c53afbb/87571a6e-9633-4d94-9cec-fbb68c53afbb.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1102.917531] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1102.917791] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5deac20c-83b3-4241-b26f-ec37f26fb4de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.924657] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1102.924657] env[63197]: value = "task-1364671" [ 1102.924657] env[63197]: _type = "Task" [ 1102.924657] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.934227] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364671, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.433845] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364671, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.26741} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.434329] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1103.435153] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964b870d-402d-4226-8537-aba422a958f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.457154] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 87571a6e-9633-4d94-9cec-fbb68c53afbb/87571a6e-9633-4d94-9cec-fbb68c53afbb.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1103.457431] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f1b7e19-0504-4d77-b5a1-748a0e29a7ea {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.475966] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1103.475966] env[63197]: value = "task-1364672" [ 1103.475966] env[63197]: _type = "Task" [ 1103.475966] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.483458] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364672, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.656513] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.656717] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.986577] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.159436] env[63197]: DEBUG nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1104.486667] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364672, 'name': ReconfigVM_Task, 'duration_secs': 0.585739} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.487022] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 87571a6e-9633-4d94-9cec-fbb68c53afbb/87571a6e-9633-4d94-9cec-fbb68c53afbb.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1104.487616] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7dd8265-377e-4d27-94e0-313aa4ad9a31 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.495455] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1104.495455] env[63197]: value = "task-1364674" [ 1104.495455] env[63197]: _type = "Task" [ 1104.495455] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.506795] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364674, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.681825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.682192] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.683695] env[63197]: INFO nova.compute.claims [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1105.005545] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364674, 'name': Rename_Task, 'duration_secs': 0.169299} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.005824] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1105.006083] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a89bd6c6-9c7a-40aa-a88e-26dedb8051c0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.012731] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1105.012731] env[63197]: value = "task-1364675" [ 1105.012731] env[63197]: _type = "Task" [ 1105.012731] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.019720] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364675, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.523275] env[63197]: DEBUG oslo_vmware.api [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364675, 'name': PowerOnVM_Task, 'duration_secs': 0.435094} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.523570] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1105.523693] env[63197]: INFO nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1105.523873] env[63197]: DEBUG nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1105.524653] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86453311-2613-48d7-8182-79a71fac9c39 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.739773] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b525053-043b-4594-a125-31d226df11aa {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.748061] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b72e9bc-5442-4585-90ef-19f41200ac54 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.776413] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4213538d-fc34-4172-820f-5c5739be78ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.783325] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4960cdba-bdc2-42c9-bdf2-7dae71ab77c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.797511] env[63197]: DEBUG nova.compute.provider_tree [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.041651] env[63197]: INFO nova.compute.manager [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Took 12.40 seconds to build instance. [ 1106.300838] env[63197]: DEBUG nova.scheduler.client.report [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1106.544060] env[63197]: DEBUG oslo_concurrency.lockutils [None req-7f2818da-6835-422e-8135-09fb8f5e3831 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.926s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.625336] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.625559] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.625751] env[63197]: INFO nova.compute.manager [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Shelving [ 1106.805622] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.123s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.806134] env[63197]: DEBUG nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1106.916612] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.916861] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.917085] env[63197]: DEBUG nova.compute.manager [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1106.918015] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04eee01-8008-46c9-83e2-562c7f43b445 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.925385] env[63197]: DEBUG nova.compute.manager [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63197) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1106.925962] env[63197]: DEBUG nova.objects.instance [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'flavor' on Instance uuid 87571a6e-9633-4d94-9cec-fbb68c53afbb {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.132846] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1107.133132] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b40feb2e-5aba-41b5-8a71-6463a517eb83 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.142415] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1107.142415] env[63197]: value = "task-1364676" [ 1107.142415] env[63197]: _type = "Task" [ 1107.142415] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.154383] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.311061] env[63197]: DEBUG nova.compute.utils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1107.312599] env[63197]: DEBUG nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1107.312765] env[63197]: DEBUG nova.network.neutron [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1107.357632] env[63197]: DEBUG nova.policy [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '514b7795ecdc49069e7f3cab700a0ebd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '527721557235413e99c4a2eaa086486c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1107.430613] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1107.430895] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8bbaafd3-fe2d-4c9c-8bb1-84d0b61783d3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.438685] env[63197]: DEBUG oslo_vmware.api [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1107.438685] env[63197]: value = "task-1364677" [ 1107.438685] env[63197]: _type = "Task" [ 1107.438685] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.448797] env[63197]: DEBUG oslo_vmware.api [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.653891] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364676, 'name': PowerOffVM_Task, 'duration_secs': 0.174381} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.654270] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1107.655534] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb1409a-af63-4346-af2b-437b3a243d58 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.675452] env[63197]: DEBUG nova.network.neutron [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Successfully created port: 7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1107.677817] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20684dd-0b23-4811-aa9a-90b37941c1c2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.713159] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquiring lock "14456558-a814-45cd-9596-0c073c581b57" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.713479] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "14456558-a814-45cd-9596-0c073c581b57" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.815931] env[63197]: DEBUG nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1107.948779] env[63197]: DEBUG oslo_vmware.api [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364677, 'name': PowerOffVM_Task, 'duration_secs': 0.186445} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.949071] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1107.949262] env[63197]: DEBUG nova.compute.manager [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1107.950015] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62330320-1f25-45ac-851c-e2db78c9331a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.190268] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Creating Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1108.190624] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8e79923c-ecba-44eb-9a95-af6eb0b3e0d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.199769] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1108.199769] env[63197]: value = "task-1364678" [ 1108.199769] env[63197]: _type = "Task" [ 1108.199769] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.208407] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364678, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.216014] env[63197]: DEBUG nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1108.460909] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bedd428f-d30a-4d72-8665-776c5ee1b9b7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.712449] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364678, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.739236] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.739539] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.741256] env[63197]: INFO nova.compute.claims [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1108.826419] env[63197]: DEBUG nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1108.855640] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.855948] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.856147] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.856344] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.856489] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.856662] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.856897] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.857086] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.857294] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.857570] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.857800] env[63197]: DEBUG nova.virt.hardware [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.859097] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e27cdde2-bb18-4cdb-9530-e7a4a28200f4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.868507] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8568060a-7b93-42d2-9abd-d3ce1c9ea807 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.061311] env[63197]: DEBUG nova.compute.manager [req-3ec7a469-6579-40b8-a7b5-4d38e9847549 req-5e02ed5f-28c2-42b8-9175-f900abc91948 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Received event network-vif-plugged-7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.061550] env[63197]: DEBUG oslo_concurrency.lockutils [req-3ec7a469-6579-40b8-a7b5-4d38e9847549 req-5e02ed5f-28c2-42b8-9175-f900abc91948 service nova] Acquiring lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.061769] env[63197]: DEBUG oslo_concurrency.lockutils [req-3ec7a469-6579-40b8-a7b5-4d38e9847549 req-5e02ed5f-28c2-42b8-9175-f900abc91948 service nova] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.061940] env[63197]: DEBUG oslo_concurrency.lockutils [req-3ec7a469-6579-40b8-a7b5-4d38e9847549 req-5e02ed5f-28c2-42b8-9175-f900abc91948 service nova] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.062228] env[63197]: DEBUG nova.compute.manager [req-3ec7a469-6579-40b8-a7b5-4d38e9847549 req-5e02ed5f-28c2-42b8-9175-f900abc91948 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] No waiting events found dispatching network-vif-plugged-7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1109.062403] env[63197]: WARNING nova.compute.manager [req-3ec7a469-6579-40b8-a7b5-4d38e9847549 req-5e02ed5f-28c2-42b8-9175-f900abc91948 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Received unexpected event network-vif-plugged-7b0c52af-a5fc-4266-8dd0-73c379e73cdd for instance with vm_state building and task_state spawning. [ 1109.146371] env[63197]: DEBUG nova.network.neutron [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Successfully updated port: 7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1109.210526] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364678, 'name': CreateSnapshot_Task, 'duration_secs': 0.5736} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.210832] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Created Snapshot of the VM instance {{(pid=63197) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1109.211659] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914ff40d-d43c-41b1-a160-50216bd34d37 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.310536] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.310806] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.311033] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "87571a6e-9633-4d94-9cec-fbb68c53afbb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.311245] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.311403] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.313422] env[63197]: INFO nova.compute.manager [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Terminating instance [ 1109.315175] env[63197]: DEBUG nova.compute.manager [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1109.315376] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1109.316188] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216c1f75-8e29-47cf-9c35-f000a5fbe223 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.324337] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.324558] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4032bde-13a8-4a01-90d1-ac9225fb97f8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.387570] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.387808] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.387992] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleting the datastore file [datastore2] 87571a6e-9633-4d94-9cec-fbb68c53afbb {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.388302] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b7ccf74-aa58-4e46-acc3-8273a80fc2a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.395085] env[63197]: DEBUG oslo_vmware.api [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1109.395085] env[63197]: value = "task-1364680" [ 1109.395085] env[63197]: _type = "Task" [ 1109.395085] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.402731] env[63197]: DEBUG oslo_vmware.api [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364680, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.648777] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.648934] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.649131] env[63197]: DEBUG nova.network.neutron [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1109.730034] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Creating linked-clone VM from snapshot {{(pid=63197) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1109.730629] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a5204644-6572-4857-974b-b360213ac40e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.739582] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1109.739582] env[63197]: value = "task-1364681" [ 1109.739582] env[63197]: _type = "Task" [ 1109.739582] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.751486] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364681, 'name': CloneVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.815657] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec61da78-539f-4a9a-a35c-dbc52774c72d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.823413] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a318133-b3c5-476e-8595-0511ef442616 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.855180] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3f4506-dd28-4868-a2d9-a165f3b2cd4a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.862868] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a8bd71-3ec5-46d2-b525-0f35c38589cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.877072] env[63197]: DEBUG nova.compute.provider_tree [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.904533] env[63197]: DEBUG oslo_vmware.api [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364680, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128558} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.904795] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.904987] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1109.905212] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.905395] env[63197]: INFO nova.compute.manager [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1109.905649] env[63197]: DEBUG oslo.service.loopingcall [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1109.905871] env[63197]: DEBUG nova.compute.manager [-] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1109.905963] env[63197]: DEBUG nova.network.neutron [-] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1110.194665] env[63197]: DEBUG nova.network.neutron [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1110.250259] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364681, 'name': CloneVM_Task} progress is 94%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.322370] env[63197]: DEBUG nova.network.neutron [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updating instance_info_cache with network_info: [{"id": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "address": "fa:16:3e:db:00:06", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b0c52af-a5", "ovs_interfaceid": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.380638] env[63197]: DEBUG nova.scheduler.client.report [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1110.682779] env[63197]: DEBUG nova.network.neutron [-] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.751070] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364681, 'name': CloneVM_Task, 'duration_secs': 0.932374} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.751368] env[63197]: INFO nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Created linked-clone VM from snapshot [ 1110.752111] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc43d0f-63c5-4095-8776-1c523b117307 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.759472] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Uploading image 967e0b8c-0df0-4299-8693-ceb1aebdc831 {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1110.787152] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1110.787152] env[63197]: value = "vm-290424" [ 1110.787152] env[63197]: _type = "VirtualMachine" [ 1110.787152] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1110.787320] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-59befdc8-b26e-4cd0-a15c-d2f0c77a228c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.795979] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease: (returnval){ [ 1110.795979] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52976344-963f-fd04-a3f3-d5860799cf0e" [ 1110.795979] env[63197]: _type = "HttpNfcLease" [ 1110.795979] env[63197]: } obtained for exporting VM: (result){ [ 1110.795979] env[63197]: value = "vm-290424" [ 1110.795979] env[63197]: _type = "VirtualMachine" [ 1110.795979] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1110.796278] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the lease: (returnval){ [ 1110.796278] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52976344-963f-fd04-a3f3-d5860799cf0e" [ 1110.796278] env[63197]: _type = "HttpNfcLease" [ 1110.796278] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1110.802765] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1110.802765] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52976344-963f-fd04-a3f3-d5860799cf0e" [ 1110.802765] env[63197]: _type = "HttpNfcLease" [ 1110.802765] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1110.825414] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.825730] env[63197]: DEBUG nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Instance network_info: |[{"id": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "address": "fa:16:3e:db:00:06", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b0c52af-a5", "ovs_interfaceid": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1110.826476] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:00:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa8c2f93-f287-41b3-adb6-4942a7ea2a0b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b0c52af-a5fc-4266-8dd0-73c379e73cdd', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.834398] env[63197]: DEBUG oslo.service.loopingcall [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.834622] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1110.834846] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37fc2401-c77b-496d-af70-887008f0762d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.855592] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.855592] env[63197]: value = "task-1364683" [ 1110.855592] env[63197]: _type = "Task" [ 1110.855592] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.863374] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364683, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.885285] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.146s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.885660] env[63197]: DEBUG nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1111.088214] env[63197]: DEBUG nova.compute.manager [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Received event network-changed-7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1111.088427] env[63197]: DEBUG nova.compute.manager [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Refreshing instance network info cache due to event network-changed-7b0c52af-a5fc-4266-8dd0-73c379e73cdd. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1111.088652] env[63197]: DEBUG oslo_concurrency.lockutils [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] Acquiring lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.088801] env[63197]: DEBUG oslo_concurrency.lockutils [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] Acquired lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.088962] env[63197]: DEBUG nova.network.neutron [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Refreshing network info cache for port 7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1111.185684] env[63197]: INFO nova.compute.manager [-] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Took 1.28 seconds to deallocate network for instance. [ 1111.305066] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1111.305066] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52976344-963f-fd04-a3f3-d5860799cf0e" [ 1111.305066] env[63197]: _type = "HttpNfcLease" [ 1111.305066] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1111.305372] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1111.305372] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52976344-963f-fd04-a3f3-d5860799cf0e" [ 1111.305372] env[63197]: _type = "HttpNfcLease" [ 1111.305372] env[63197]: }. {{(pid=63197) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1111.306283] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c33d3a69-6a84-4bbf-89f6-9cd2b140330a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.313164] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e5b5e9-276f-4a4f-c36c-42efb9d4719a/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1111.313343] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e5b5e9-276f-4a4f-c36c-42efb9d4719a/disk-0.vmdk for reading. {{(pid=63197) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1111.380968] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364683, 'name': CreateVM_Task, 'duration_secs': 0.390558} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.381162] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1111.381841] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.382023] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.382346] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1111.382611] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bd02db0-445d-4667-8e9d-8a073e164ace {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.387460] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1111.387460] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52955e5d-0fe8-7491-b772-a9dc24a1983e" [ 1111.387460] env[63197]: _type = "Task" [ 1111.387460] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.391291] env[63197]: DEBUG nova.compute.utils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1111.393112] env[63197]: DEBUG nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1111.393299] env[63197]: DEBUG nova.network.neutron [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1111.401407] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52955e5d-0fe8-7491-b772-a9dc24a1983e, 'name': SearchDatastore_Task, 'duration_secs': 0.009104} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.402501] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.402741] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1111.402996] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.403174] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquired lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.403564] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1111.406456] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ac8803e-bc0a-47f7-990d-251a39378d97 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.409768] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fcf6249a-0b47-4b98-8b98-4fcd15f6b509 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.418545] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1111.418737] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1111.419464] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1aac034e-4de3-41a4-a6d1-59e25a8ee6be {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.427096] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1111.427096] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a2c740-386c-d2d7-99c6-8804e34301de" [ 1111.427096] env[63197]: _type = "Task" [ 1111.427096] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.438908] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a2c740-386c-d2d7-99c6-8804e34301de, 'name': SearchDatastore_Task, 'duration_secs': 0.009139} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.445565] env[63197]: DEBUG nova.policy [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '29ca771c87e44eb6aa09d6b28cb311f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '168104be9877419e896624acc5bdcdc5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1111.447115] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50880d8d-dba4-48f7-9496-c605ca30391d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.452571] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1111.452571] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f26e6-69d8-a047-e7d3-ab96b6260fd6" [ 1111.452571] env[63197]: _type = "Task" [ 1111.452571] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.462481] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f26e6-69d8-a047-e7d3-ab96b6260fd6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.698100] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.698386] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.698618] env[63197]: DEBUG nova.objects.instance [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'resources' on Instance uuid 87571a6e-9633-4d94-9cec-fbb68c53afbb {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.894952] env[63197]: DEBUG nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1111.965830] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520f26e6-69d8-a047-e7d3-ab96b6260fd6, 'name': SearchDatastore_Task, 'duration_secs': 0.009316} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.968413] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Releasing lock "[datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.970026] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 0dc57420-b13f-4305-ae1f-6ebf61f356f3/0dc57420-b13f-4305-ae1f-6ebf61f356f3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1111.970723] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f14b66e7-1b1a-46dd-8ded-3a9417e1fcb0 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.983142] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1111.983142] env[63197]: value = "task-1364684" [ 1111.983142] env[63197]: _type = "Task" [ 1111.983142] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.998083] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364684, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.073226] env[63197]: DEBUG nova.network.neutron [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Successfully created port: f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1112.204381] env[63197]: DEBUG nova.network.neutron [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updated VIF entry in instance network info cache for port 7b0c52af-a5fc-4266-8dd0-73c379e73cdd. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1112.204730] env[63197]: DEBUG nova.network.neutron [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updating instance_info_cache with network_info: [{"id": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "address": "fa:16:3e:db:00:06", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b0c52af-a5", "ovs_interfaceid": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.285967] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38ede0d-1a45-4988-9899-76ebb4af5068 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.294840] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae82447-302b-445b-992f-57f20d77c36a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.328899] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3e0627-ee3c-437e-86d3-f19c8b4cd36f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.338896] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb2f854-6242-4883-a399-1d7442605b1a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.358061] env[63197]: DEBUG nova.compute.provider_tree [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.400419] env[63197]: INFO nova.virt.block_device [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Booting with volume 5960526b-c5d9-4c8b-ad0b-ba59179fa82a at /dev/sda [ 1112.448037] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d081b414-6a00-431a-832a-287ad3bee9d5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.458021] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddc5644-38ff-4640-99d8-50da9e83e49e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.490979] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eca12d78-c9b7-45fd-9652-9d2b12ee2740 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.500753] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364684, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.505551] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3929c5-08cd-4d80-8a51-332a45db9098 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.536330] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8ea69d-0afc-4dda-8e6b-d27ad2c60596 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.545344] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac537a9-930a-46cd-ac2f-e35f442816c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.565060] env[63197]: DEBUG nova.virt.block_device [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updating existing volume attachment record: 98463045-bb9f-4df1-918d-c7c880178139 {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1112.712205] env[63197]: DEBUG oslo_concurrency.lockutils [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] Releasing lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.712497] env[63197]: DEBUG nova.compute.manager [req-ca2a6438-e8d1-4523-9ca8-778f3ece16a1 req-0471f6b3-6083-43e9-9f34-9936df6174eb service nova] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Received event network-vif-deleted-c0f5ed33-2425-4178-82ea-7eadf90cc2cf {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1112.860048] env[63197]: DEBUG nova.scheduler.client.report [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1112.998033] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364684, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707581} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.998427] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore1] 0dc57420-b13f-4305-ae1f-6ebf61f356f3/0dc57420-b13f-4305-ae1f-6ebf61f356f3.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1112.998758] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1112.999097] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb5c4333-feb3-45f5-99e4-638c53f47db7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.007119] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1113.007119] env[63197]: value = "task-1364685" [ 1113.007119] env[63197]: _type = "Task" [ 1113.007119] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.017460] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364685, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.364928] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.385350] env[63197]: INFO nova.scheduler.client.report [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted allocations for instance 87571a6e-9633-4d94-9cec-fbb68c53afbb [ 1113.517834] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364685, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066042} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.518191] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1113.519026] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-340def61-b72a-4977-bfee-7ebb81ba32d6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.543797] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 0dc57420-b13f-4305-ae1f-6ebf61f356f3/0dc57420-b13f-4305-ae1f-6ebf61f356f3.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.544208] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0aa00983-c439-4ee8-825b-6a2911f78122 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.560509] env[63197]: DEBUG nova.compute.manager [req-172eaa94-4289-4be5-ac42-dd2441864706 req-a0c3f026-652c-4d1a-a98d-d4c34c76c77e service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Received event network-vif-plugged-f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1113.560985] env[63197]: DEBUG oslo_concurrency.lockutils [req-172eaa94-4289-4be5-ac42-dd2441864706 req-a0c3f026-652c-4d1a-a98d-d4c34c76c77e service nova] Acquiring lock "14456558-a814-45cd-9596-0c073c581b57-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.561234] env[63197]: DEBUG oslo_concurrency.lockutils [req-172eaa94-4289-4be5-ac42-dd2441864706 req-a0c3f026-652c-4d1a-a98d-d4c34c76c77e service nova] Lock "14456558-a814-45cd-9596-0c073c581b57-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.561535] env[63197]: DEBUG oslo_concurrency.lockutils [req-172eaa94-4289-4be5-ac42-dd2441864706 req-a0c3f026-652c-4d1a-a98d-d4c34c76c77e service nova] Lock "14456558-a814-45cd-9596-0c073c581b57-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.561801] env[63197]: DEBUG nova.compute.manager [req-172eaa94-4289-4be5-ac42-dd2441864706 req-a0c3f026-652c-4d1a-a98d-d4c34c76c77e service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] No waiting events found dispatching network-vif-plugged-f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1113.562062] env[63197]: WARNING nova.compute.manager [req-172eaa94-4289-4be5-ac42-dd2441864706 req-a0c3f026-652c-4d1a-a98d-d4c34c76c77e service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Received unexpected event network-vif-plugged-f3255d63-3c6b-4d3e-9134-4ab33d8da179 for instance with vm_state building and task_state block_device_mapping. [ 1113.572027] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1113.572027] env[63197]: value = "task-1364686" [ 1113.572027] env[63197]: _type = "Task" [ 1113.572027] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.578730] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364686, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.648812] env[63197]: DEBUG nova.network.neutron [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Successfully updated port: f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1113.894092] env[63197]: DEBUG oslo_concurrency.lockutils [None req-4b8604d6-b239-4c4a-a1f1-0a993c028e4a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "87571a6e-9633-4d94-9cec-fbb68c53afbb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.582s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.080403] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364686, 'name': ReconfigVM_Task, 'duration_secs': 0.334577} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.081138] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 0dc57420-b13f-4305-ae1f-6ebf61f356f3/0dc57420-b13f-4305-ae1f-6ebf61f356f3.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.081819] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69b0e78b-38ea-452c-970a-6764f039f64f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.090472] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1114.090472] env[63197]: value = "task-1364687" [ 1114.090472] env[63197]: _type = "Task" [ 1114.090472] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.100273] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364687, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.152248] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquiring lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.152472] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquired lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.152690] env[63197]: DEBUG nova.network.neutron [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1114.602176] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364687, 'name': Rename_Task, 'duration_secs': 0.138306} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.602493] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1114.602750] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed27c1b5-cb70-41ec-b91f-5ea20b36d24e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.612713] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1114.612713] env[63197]: value = "task-1364688" [ 1114.612713] env[63197]: _type = "Task" [ 1114.612713] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.622625] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364688, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.666243] env[63197]: DEBUG nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1114.666833] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1114.667619] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1114.667619] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1114.667713] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1114.667926] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1114.668185] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1114.668654] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1114.668719] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1114.668946] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1114.669289] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1114.669494] env[63197]: DEBUG nova.virt.hardware [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1114.670541] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69aafcd-8539-45f8-bffb-8a0a1b4c7e30 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.680204] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1948b4-1c22-4059-bf43-983324acc3a9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.685889] env[63197]: DEBUG nova.network.neutron [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1114.863125] env[63197]: DEBUG nova.network.neutron [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updating instance_info_cache with network_info: [{"id": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "address": "fa:16:3e:a0:de:c5", "network": {"id": "acdfcb1f-bde5-4cf6-831d-33c5fca79249", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1884926183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "168104be9877419e896624acc5bdcdc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3255d63-3c", "ovs_interfaceid": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.125190] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364688, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.216680] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "95e316b6-1162-414f-a6ca-43b37eb6597c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.216898] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.366664] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Releasing lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.367011] env[63197]: DEBUG nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Instance network_info: |[{"id": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "address": "fa:16:3e:a0:de:c5", "network": {"id": "acdfcb1f-bde5-4cf6-831d-33c5fca79249", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1884926183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "168104be9877419e896624acc5bdcdc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3255d63-3c", "ovs_interfaceid": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1115.367512] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:de:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4345ef6-a7c8-4c1c-badf-a0d4f578b61c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3255d63-3c6b-4d3e-9134-4ab33d8da179', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1115.376115] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Creating folder: Project (168104be9877419e896624acc5bdcdc5). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1115.376468] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-439df493-d2b0-4b84-ac2f-8739ea25039b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.392022] env[63197]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1115.392259] env[63197]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63197) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1115.392699] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Folder already exists: Project (168104be9877419e896624acc5bdcdc5). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1115.392963] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Creating folder: Instances. Parent ref: group-v290418. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1115.393272] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1413ffdc-7803-437c-886d-77e7ab7f19a2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.407474] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Created folder: Instances in parent group-v290418. [ 1115.407742] env[63197]: DEBUG oslo.service.loopingcall [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.407955] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14456558-a814-45cd-9596-0c073c581b57] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1115.408194] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1173249d-208b-4899-bfb7-c2be4b629bc1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.429144] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1115.429144] env[63197]: value = "task-1364691" [ 1115.429144] env[63197]: _type = "Task" [ 1115.429144] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.442396] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364691, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.577685] env[63197]: DEBUG nova.compute.manager [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Received event network-changed-f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.577892] env[63197]: DEBUG nova.compute.manager [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Refreshing instance network info cache due to event network-changed-f3255d63-3c6b-4d3e-9134-4ab33d8da179. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1115.578137] env[63197]: DEBUG oslo_concurrency.lockutils [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] Acquiring lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.578289] env[63197]: DEBUG oslo_concurrency.lockutils [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] Acquired lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.578451] env[63197]: DEBUG nova.network.neutron [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Refreshing network info cache for port f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1115.624612] env[63197]: DEBUG oslo_vmware.api [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364688, 'name': PowerOnVM_Task, 'duration_secs': 0.535042} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.624870] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1115.625110] env[63197]: INFO nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Took 6.80 seconds to spawn the instance on the hypervisor. [ 1115.625316] env[63197]: DEBUG nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.626168] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424fc7e8-6c2b-4664-a387-b56447e570bf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.719165] env[63197]: DEBUG nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1115.940565] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364691, 'name': CreateVM_Task, 'duration_secs': 0.478828} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.940746] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14456558-a814-45cd-9596-0c073c581b57] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1115.941503] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'boot_index': 0, 'disk_bus': None, 'attachment_id': '98463045-bb9f-4df1-918d-c7c880178139', 'mount_device': '/dev/sda', 'delete_on_termination': True, 'guest_format': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290421', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'name': 'volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '14456558-a814-45cd-9596-0c073c581b57', 'attached_at': '', 'detached_at': '', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'serial': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a'}, 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=63197) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1115.941756] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Root volume attach. Driver type: vmdk {{(pid=63197) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1115.942557] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84abb9b-19ff-4b0e-8a79-3e8784887946 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.951050] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6085bdd4-a645-43ee-bb8f-70aca72c031b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.957946] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1659f4ef-2116-4959-a751-63d3534eb6bc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.966596] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-8b2de8dd-d63f-4299-abfc-f19dfc246212 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.975902] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1115.975902] env[63197]: value = "task-1364692" [ 1115.975902] env[63197]: _type = "Task" [ 1115.975902] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.985710] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1115.985993] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.986222] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.143345] env[63197]: INFO nova.compute.manager [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Took 11.48 seconds to build instance. [ 1116.249764] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.250287] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.253075] env[63197]: INFO nova.compute.claims [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.340961] env[63197]: DEBUG nova.network.neutron [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updated VIF entry in instance network info cache for port f3255d63-3c6b-4d3e-9134-4ab33d8da179. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1116.340961] env[63197]: DEBUG nova.network.neutron [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updating instance_info_cache with network_info: [{"id": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "address": "fa:16:3e:a0:de:c5", "network": {"id": "acdfcb1f-bde5-4cf6-831d-33c5fca79249", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1884926183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "168104be9877419e896624acc5bdcdc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3255d63-3c", "ovs_interfaceid": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.488524] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task} progress is 42%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.491175] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.491391] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.491547] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.491689] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.491839] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.491983] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1116.492188] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1116.645484] env[63197]: DEBUG oslo_concurrency.lockutils [None req-f64141cc-59a1-402e-a442-ccd2c68079db tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.988s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.842125] env[63197]: DEBUG oslo_concurrency.lockutils [req-57b11782-85d0-40a8-b9a2-f1d563f43453 req-70d9930d-c34e-41ba-9620-5de4481334a8 service nova] Releasing lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.990644] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task} progress is 54%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.995042] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.203320] env[63197]: DEBUG nova.compute.manager [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Received event network-changed-7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1117.203612] env[63197]: DEBUG nova.compute.manager [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Refreshing instance network info cache due to event network-changed-7b0c52af-a5fc-4266-8dd0-73c379e73cdd. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1117.203659] env[63197]: DEBUG oslo_concurrency.lockutils [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] Acquiring lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1117.203806] env[63197]: DEBUG oslo_concurrency.lockutils [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] Acquired lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.203973] env[63197]: DEBUG nova.network.neutron [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Refreshing network info cache for port 7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1117.346266] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef63dfd0-4ef3-417e-9e6c-6cc96974f460 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.357371] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8484d007-b80f-41a1-998b-cea4cab390cf {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.396293] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e3ee69-55e9-4aa9-ac19-e21fa61aa99f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.406901] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724cddb3-388f-461e-96ab-adf0c065d964 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.427447] env[63197]: DEBUG nova.compute.provider_tree [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.491671] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task} progress is 67%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.930538] env[63197]: DEBUG nova.scheduler.client.report [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.977892] env[63197]: DEBUG nova.network.neutron [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updated VIF entry in instance network info cache for port 7b0c52af-a5fc-4266-8dd0-73c379e73cdd. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1117.978309] env[63197]: DEBUG nova.network.neutron [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updating instance_info_cache with network_info: [{"id": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "address": "fa:16:3e:db:00:06", "network": {"id": "4887d49c-2a0b-47f3-a180-6535b7955544", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1366265100-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "527721557235413e99c4a2eaa086486c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa8c2f93-f287-41b3-adb6-4942a7ea2a0b", "external-id": "nsx-vlan-transportzone-363", "segmentation_id": 363, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b0c52af-a5", "ovs_interfaceid": "7b0c52af-a5fc-4266-8dd0-73c379e73cdd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.992964] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task} progress is 81%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.437761] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.438462] env[63197]: DEBUG nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1118.441887] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.447s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.442264] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.442458] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1118.443727] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7623418-7f76-49ff-8fa6-92a7c90366ce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.455386] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8dfd78-feb9-4959-98fd-2b7bd198770b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.471071] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a0b07e-6ad2-4f41-94d7-10a8980a235a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.480015] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1b338b-5143-4269-b2a3-0d9eff0a2a2d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.484147] env[63197]: DEBUG oslo_concurrency.lockutils [req-32c602b1-0ff0-4e5a-97b3-6885c40163fe req-a565e42a-edaf-4dbe-b968-4d79859b8f18 service nova] Releasing lock "refresh_cache-0dc57420-b13f-4305-ae1f-6ebf61f356f3" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.493635] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task} progress is 95%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.526327] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180775MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1118.526606] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.527017] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.945447] env[63197]: DEBUG nova.compute.utils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1118.947681] env[63197]: DEBUG nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1118.947876] env[63197]: DEBUG nova.network.neutron [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1118.990924] env[63197]: DEBUG nova.policy [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a8e87b64fe46d7ab7e570d57611119', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ceb1016d6d34bff8880dca42d495377', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1118.999789] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task} progress is 97%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.258953] env[63197]: DEBUG nova.network.neutron [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Successfully created port: a208670d-3e72-4aef-a6bb-8aba06a94df2 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1119.452497] env[63197]: DEBUG nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1119.496280] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364692, 'name': RelocateVM_Task, 'duration_secs': 3.386237} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.496280] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1119.496280] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290421', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'name': 'volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '14456558-a814-45cd-9596-0c073c581b57', 'attached_at': '', 'detached_at': '', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'serial': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1119.497447] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541100ff-2a73-4a44-87ba-d72a86c9c03a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.515727] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8433cbbb-64d0-4ed4-a971-ce7493f8f095 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.539290] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a/volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.542172] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90bc1b81-7ce5-421a-9d33-546d2cdef349 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.564556] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1119.564556] env[63197]: value = "task-1364693" [ 1119.564556] env[63197]: _type = "Task" [ 1119.564556] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.574259] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364693, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.575254] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance ad216370-3f6d-4ebe-a956-7dd3bf05511f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1119.575395] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 0dc57420-b13f-4305-ae1f-6ebf61f356f3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1119.575518] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 14456558-a814-45cd-9596-0c073c581b57 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1119.575644] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 95e316b6-1162-414f-a6ca-43b37eb6597c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1119.575811] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1119.575947] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1119.639244] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4b9218-ea20-4850-9132-bef39e4e503a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.647347] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34904137-36e5-48a8-ad0e-98dec6ae68bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.680516] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f58918b-aa0b-4f78-b06a-55c097b1a0c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.689403] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4b2b0ad-0203-437f-8908-b71a3680fa40 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.704631] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.075536] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364693, 'name': ReconfigVM_Task, 'duration_secs': 0.413468} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.075867] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Reconfigured VM instance instance-00000067 to attach disk [datastore2] volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a/volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.080442] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f682eb66-32fe-4f3c-a667-4c8a3de8b966 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.097313] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1120.097313] env[63197]: value = "task-1364694" [ 1120.097313] env[63197]: _type = "Task" [ 1120.097313] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.106133] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364694, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.208175] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.463167] env[63197]: DEBUG nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1120.502931] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1120.503431] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1120.503760] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1120.504169] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1120.504540] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1120.504891] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1120.505260] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1120.505480] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1120.505703] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1120.505944] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1120.506198] env[63197]: DEBUG nova.virt.hardware [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1120.507170] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94bea6a-8d82-4598-983f-64aa97855e86 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.515919] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a69fbf1-50d9-47f2-98f2-3eb565340a6f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.608620] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364694, 'name': ReconfigVM_Task, 'duration_secs': 0.142494} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.608996] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290421', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'name': 'volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '14456558-a814-45cd-9596-0c073c581b57', 'attached_at': '', 'detached_at': '', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'serial': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1120.610037] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16eda49d-799c-4979-a141-5e8ddaad1cab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.617906] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1120.617906] env[63197]: value = "task-1364695" [ 1120.617906] env[63197]: _type = "Task" [ 1120.617906] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.630348] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364695, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.673586] env[63197]: DEBUG nova.compute.manager [req-0879d9d7-8324-4c73-9b45-46fabfe5b01b req-76c9856b-02a5-4894-82b9-522cea6136a7 service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Received event network-vif-plugged-a208670d-3e72-4aef-a6bb-8aba06a94df2 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1120.673832] env[63197]: DEBUG oslo_concurrency.lockutils [req-0879d9d7-8324-4c73-9b45-46fabfe5b01b req-76c9856b-02a5-4894-82b9-522cea6136a7 service nova] Acquiring lock "95e316b6-1162-414f-a6ca-43b37eb6597c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.674046] env[63197]: DEBUG oslo_concurrency.lockutils [req-0879d9d7-8324-4c73-9b45-46fabfe5b01b req-76c9856b-02a5-4894-82b9-522cea6136a7 service nova] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.674337] env[63197]: DEBUG oslo_concurrency.lockutils [req-0879d9d7-8324-4c73-9b45-46fabfe5b01b req-76c9856b-02a5-4894-82b9-522cea6136a7 service nova] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.674565] env[63197]: DEBUG nova.compute.manager [req-0879d9d7-8324-4c73-9b45-46fabfe5b01b req-76c9856b-02a5-4894-82b9-522cea6136a7 service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] No waiting events found dispatching network-vif-plugged-a208670d-3e72-4aef-a6bb-8aba06a94df2 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1120.674811] env[63197]: WARNING nova.compute.manager [req-0879d9d7-8324-4c73-9b45-46fabfe5b01b req-76c9856b-02a5-4894-82b9-522cea6136a7 service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Received unexpected event network-vif-plugged-a208670d-3e72-4aef-a6bb-8aba06a94df2 for instance with vm_state building and task_state spawning. [ 1120.712890] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1120.713207] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.186s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.713445] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.713645] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Cleaning up deleted instances {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1120.765532] env[63197]: DEBUG nova.network.neutron [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Successfully updated port: a208670d-3e72-4aef-a6bb-8aba06a94df2 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1121.129503] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364695, 'name': Rename_Task, 'duration_secs': 0.172542} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.129612] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1121.129816] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b959920-6ac9-43c0-a430-84e5dfaa3543 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.136960] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1121.136960] env[63197]: value = "task-1364696" [ 1121.136960] env[63197]: _type = "Task" [ 1121.136960] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.146128] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.165025] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e5b5e9-276f-4a4f-c36c-42efb9d4719a/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1121.165572] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcb5070-b6b6-4133-8d20-eb3212d2e19f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.174118] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e5b5e9-276f-4a4f-c36c-42efb9d4719a/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1121.175099] env[63197]: ERROR oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e5b5e9-276f-4a4f-c36c-42efb9d4719a/disk-0.vmdk due to incomplete transfer. [ 1121.175099] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ce4967f9-af50-45e1-8206-6d08c530fa21 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.181663] env[63197]: DEBUG oslo_vmware.rw_handles [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e5b5e9-276f-4a4f-c36c-42efb9d4719a/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1121.182079] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Uploaded image 967e0b8c-0df0-4299-8693-ceb1aebdc831 to the Glance image server {{(pid=63197) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1121.184456] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Destroying the VM {{(pid=63197) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1121.184803] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c5caa936-5c95-4f53-ab2e-921d490056cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.191333] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1121.191333] env[63197]: value = "task-1364697" [ 1121.191333] env[63197]: _type = "Task" [ 1121.191333] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.199686] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364697, 'name': Destroy_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.225807] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] There are 40 instances to clean {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1121.226208] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 87571a6e-9633-4d94-9cec-fbb68c53afbb] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.268365] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-95e316b6-1162-414f-a6ca-43b37eb6597c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.268665] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-95e316b6-1162-414f-a6ca-43b37eb6597c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.268884] env[63197]: DEBUG nova.network.neutron [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1121.647160] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364696, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.701835] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364697, 'name': Destroy_Task, 'duration_secs': 0.484426} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.702259] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Destroyed the VM [ 1121.702558] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Deleting Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1121.703215] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2066e7d6-720f-44f0-9d95-3d20b2ed6430 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.710599] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1121.710599] env[63197]: value = "task-1364698" [ 1121.710599] env[63197]: _type = "Task" [ 1121.710599] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.719420] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364698, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.730065] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 2ab00a06-545f-4674-b7e3-37354c7460a4] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1121.800526] env[63197]: DEBUG nova.network.neutron [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1121.939337] env[63197]: DEBUG nova.network.neutron [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Updating instance_info_cache with network_info: [{"id": "a208670d-3e72-4aef-a6bb-8aba06a94df2", "address": "fa:16:3e:c2:5f:a8", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa208670d-3e", "ovs_interfaceid": "a208670d-3e72-4aef-a6bb-8aba06a94df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.148156] env[63197]: DEBUG oslo_vmware.api [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364696, 'name': PowerOnVM_Task, 'duration_secs': 0.534009} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.148448] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1122.148707] env[63197]: INFO nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1122.148857] env[63197]: DEBUG nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1122.149639] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11fd7e78-90bc-4ed4-8fdd-1c93df33bfab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.220601] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364698, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.233489] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 1b234498-8fe0-475a-9cad-71d0be79307c] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.441993] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-95e316b6-1162-414f-a6ca-43b37eb6597c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.442317] env[63197]: DEBUG nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Instance network_info: |[{"id": "a208670d-3e72-4aef-a6bb-8aba06a94df2", "address": "fa:16:3e:c2:5f:a8", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa208670d-3e", "ovs_interfaceid": "a208670d-3e72-4aef-a6bb-8aba06a94df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1122.442764] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:5f:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a208670d-3e72-4aef-a6bb-8aba06a94df2', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1122.450262] env[63197]: DEBUG oslo.service.loopingcall [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1122.450473] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1122.450728] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41f79bc6-7365-4418-9faf-118b346ad215 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.471857] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1122.471857] env[63197]: value = "task-1364699" [ 1122.471857] env[63197]: _type = "Task" [ 1122.471857] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.479511] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364699, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.667029] env[63197]: INFO nova.compute.manager [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Took 13.95 seconds to build instance. [ 1122.704844] env[63197]: DEBUG nova.compute.manager [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Received event network-changed-a208670d-3e72-4aef-a6bb-8aba06a94df2 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1122.705068] env[63197]: DEBUG nova.compute.manager [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Refreshing instance network info cache due to event network-changed-a208670d-3e72-4aef-a6bb-8aba06a94df2. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1122.705460] env[63197]: DEBUG oslo_concurrency.lockutils [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] Acquiring lock "refresh_cache-95e316b6-1162-414f-a6ca-43b37eb6597c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.705636] env[63197]: DEBUG oslo_concurrency.lockutils [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] Acquired lock "refresh_cache-95e316b6-1162-414f-a6ca-43b37eb6597c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.705808] env[63197]: DEBUG nova.network.neutron [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Refreshing network info cache for port a208670d-3e72-4aef-a6bb-8aba06a94df2 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1122.722887] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364698, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.736437] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 18583a35-9cf4-4ae9-965e-be1fdc90efa2] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1122.981756] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364699, 'name': CreateVM_Task, 'duration_secs': 0.328339} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.981939] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1122.982671] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.982847] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.983194] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1122.983458] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00f5410b-6150-4bd6-9435-86d98c579d5a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.988393] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1122.988393] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529a33c7-81d6-7f88-65d9-5b3aabb57dea" [ 1122.988393] env[63197]: _type = "Task" [ 1122.988393] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.995934] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529a33c7-81d6-7f88-65d9-5b3aabb57dea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.169391] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ba1f27ef-4161-4427-bebd-4a557724760c tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "14456558-a814-45cd-9596-0c073c581b57" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.456s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.223504] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364698, 'name': RemoveSnapshot_Task, 'duration_secs': 1.228732} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.223724] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Deleted Snapshot of the VM instance {{(pid=63197) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1123.224011] env[63197]: DEBUG nova.compute.manager [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1123.224886] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c508d1-2422-425a-b205-16b4c6454566 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.241182] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: af31d3b0-52ed-408c-8110-c4241e204424] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.473719] env[63197]: DEBUG nova.network.neutron [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Updated VIF entry in instance network info cache for port a208670d-3e72-4aef-a6bb-8aba06a94df2. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1123.474141] env[63197]: DEBUG nova.network.neutron [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Updating instance_info_cache with network_info: [{"id": "a208670d-3e72-4aef-a6bb-8aba06a94df2", "address": "fa:16:3e:c2:5f:a8", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa208670d-3e", "ovs_interfaceid": "a208670d-3e72-4aef-a6bb-8aba06a94df2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.499716] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529a33c7-81d6-7f88-65d9-5b3aabb57dea, 'name': SearchDatastore_Task, 'duration_secs': 0.010388} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.500044] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.500284] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1123.500521] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.500672] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.500851] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1123.501130] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab21359a-d01f-44f0-a663-5eb2e068888d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.510885] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1123.511092] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1123.511824] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9056625d-3aaf-450a-8a65-97a9dd2a8f08 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.517946] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1123.517946] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520a8172-0a3f-8571-699c-ef4e0dc92b13" [ 1123.517946] env[63197]: _type = "Task" [ 1123.517946] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.526295] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520a8172-0a3f-8571-699c-ef4e0dc92b13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.744619] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 7f9ad29e-9cb9-4575-8dce-081c02767af5] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1123.746626] env[63197]: INFO nova.compute.manager [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Shelve offloading [ 1123.750794] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1123.751836] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08c14400-9376-4959-b9e0-d8f5b12c8335 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.760984] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1123.760984] env[63197]: value = "task-1364700" [ 1123.760984] env[63197]: _type = "Task" [ 1123.760984] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.769973] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364700, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.977183] env[63197]: DEBUG oslo_concurrency.lockutils [req-331335ce-09e6-452d-937a-67629ae20e99 req-46773b3a-fcf8-49dc-ae55-a83ed331b34b service nova] Releasing lock "refresh_cache-95e316b6-1162-414f-a6ca-43b37eb6597c" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.029519] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]520a8172-0a3f-8571-699c-ef4e0dc92b13, 'name': SearchDatastore_Task, 'duration_secs': 0.009722} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.030287] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22dae4c6-901d-4e07-9621-ada4ad86fda4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.036505] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1124.036505] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52caffd8-6709-ce48-15eb-6159a985edec" [ 1124.036505] env[63197]: _type = "Task" [ 1124.036505] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.043879] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52caffd8-6709-ce48-15eb-6159a985edec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.248838] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a77fe81f-fa44-4639-acbb-bf093b5d4f6c] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1124.272769] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] VM already powered off {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1124.272769] env[63197]: DEBUG nova.compute.manager [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1124.273344] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0031bbd-84fd-457d-99d7-8e541bf60339 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.280363] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.280509] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.280672] env[63197]: DEBUG nova.network.neutron [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1124.550190] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52caffd8-6709-ce48-15eb-6159a985edec, 'name': SearchDatastore_Task, 'duration_secs': 0.037568} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.550430] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.550695] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 95e316b6-1162-414f-a6ca-43b37eb6597c/95e316b6-1162-414f-a6ca-43b37eb6597c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1124.550966] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1828d8f9-0502-4caa-98bb-05a85e7a093a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.558874] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1124.558874] env[63197]: value = "task-1364701" [ 1124.558874] env[63197]: _type = "Task" [ 1124.558874] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.567777] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.732128] env[63197]: DEBUG nova.compute.manager [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Received event network-changed-f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1124.732448] env[63197]: DEBUG nova.compute.manager [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Refreshing instance network info cache due to event network-changed-f3255d63-3c6b-4d3e-9134-4ab33d8da179. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1124.733227] env[63197]: DEBUG oslo_concurrency.lockutils [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] Acquiring lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.733428] env[63197]: DEBUG oslo_concurrency.lockutils [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] Acquired lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.733607] env[63197]: DEBUG nova.network.neutron [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Refreshing network info cache for port f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1124.752222] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: d8919322-1fa7-4fc7-9380-a8b7d283b050] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.007755] env[63197]: DEBUG nova.network.neutron [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.071438] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364701, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.255432] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: cdd1136c-3a93-41e2-aa60-bb9b44a6bbe4] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.497568] env[63197]: DEBUG nova.network.neutron [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updated VIF entry in instance network info cache for port f3255d63-3c6b-4d3e-9134-4ab33d8da179. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1125.497959] env[63197]: DEBUG nova.network.neutron [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updating instance_info_cache with network_info: [{"id": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "address": "fa:16:3e:a0:de:c5", "network": {"id": "acdfcb1f-bde5-4cf6-831d-33c5fca79249", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1884926183-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.179", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "168104be9877419e896624acc5bdcdc5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4345ef6-a7c8-4c1c-badf-a0d4f578b61c", "external-id": "nsx-vlan-transportzone-677", "segmentation_id": 677, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3255d63-3c", "ovs_interfaceid": "f3255d63-3c6b-4d3e-9134-4ab33d8da179", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.510277] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.570703] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672129} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.570968] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 95e316b6-1162-414f-a6ca-43b37eb6597c/95e316b6-1162-414f-a6ca-43b37eb6597c.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1125.571204] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1125.571458] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f56b552-bb27-43d8-b4fe-4eee232ed337 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.580659] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1125.580659] env[63197]: value = "task-1364702" [ 1125.580659] env[63197]: _type = "Task" [ 1125.580659] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.591225] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364702, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.759250] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: f39f3e80-0d96-4024-aec8-16e20e9bf13f] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1125.812900] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.813803] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a42c68-611a-463e-86ce-8765bba3549d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.823415] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1125.823792] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0106c8b4-6b07-4a1d-aa47-3e8764381b2a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.956145] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1125.956397] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1125.956593] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleting the datastore file [datastore1] ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1125.956864] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a45d4949-2569-46fd-bbc6-a3178b092353 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.964603] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1125.964603] env[63197]: value = "task-1364704" [ 1125.964603] env[63197]: _type = "Task" [ 1125.964603] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.972136] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.000856] env[63197]: DEBUG oslo_concurrency.lockutils [req-9fbea560-2428-4352-9e9f-ec1134e0795d req-403415b8-f19c-4add-9b60-c59277fed176 service nova] Releasing lock "refresh_cache-14456558-a814-45cd-9596-0c073c581b57" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.091364] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364702, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093048} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.091615] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1126.092398] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbf6e01-d15c-46bd-b098-3cb6ab1ebec5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.114778] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 95e316b6-1162-414f-a6ca-43b37eb6597c/95e316b6-1162-414f-a6ca-43b37eb6597c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1126.115043] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7053f529-917f-43f5-a2db-e403598c163d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.134034] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1126.134034] env[63197]: value = "task-1364705" [ 1126.134034] env[63197]: _type = "Task" [ 1126.134034] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.141572] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364705, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.262435] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 430a73fa-ae22-45dd-a5c7-8c5ed616ebaa] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.474648] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364704, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.643838] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364705, 'name': ReconfigVM_Task, 'duration_secs': 0.288593} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.644143] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 95e316b6-1162-414f-a6ca-43b37eb6597c/95e316b6-1162-414f-a6ca-43b37eb6597c.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1126.644813] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0002eaaa-27ba-4b19-a4e8-e81b53a80fb5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.651980] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1126.651980] env[63197]: value = "task-1364706" [ 1126.651980] env[63197]: _type = "Task" [ 1126.651980] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.660202] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364706, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.758080] env[63197]: DEBUG nova.compute.manager [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-vif-unplugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1126.758376] env[63197]: DEBUG oslo_concurrency.lockutils [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.758613] env[63197]: DEBUG oslo_concurrency.lockutils [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.758788] env[63197]: DEBUG oslo_concurrency.lockutils [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.758960] env[63197]: DEBUG nova.compute.manager [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] No waiting events found dispatching network-vif-unplugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1126.759183] env[63197]: WARNING nova.compute.manager [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received unexpected event network-vif-unplugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af for instance with vm_state shelved and task_state shelving_offloading. [ 1126.759369] env[63197]: DEBUG nova.compute.manager [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1126.759603] env[63197]: DEBUG nova.compute.manager [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing instance network info cache due to event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1126.759716] env[63197]: DEBUG oslo_concurrency.lockutils [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.759875] env[63197]: DEBUG oslo_concurrency.lockutils [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.760037] env[63197]: DEBUG nova.network.neutron [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1126.766100] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 1bcfd6fb-d010-4eef-8096-ef9ce14d6a5f] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1126.975093] env[63197]: DEBUG oslo_vmware.api [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364704, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.650707} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.975364] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.975556] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1126.975732] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1127.069771] env[63197]: INFO nova.scheduler.client.report [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted allocations for instance ad216370-3f6d-4ebe-a956-7dd3bf05511f [ 1127.162144] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364706, 'name': Rename_Task, 'duration_secs': 0.136051} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.162410] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1127.162651] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c253011-a97e-494b-83de-9ead2ffec551 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.168958] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1127.168958] env[63197]: value = "task-1364707" [ 1127.168958] env[63197]: _type = "Task" [ 1127.168958] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.176310] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364707, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.269380] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 64b944ea-2397-45eb-a0ed-9f57263db998] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.478276] env[63197]: DEBUG nova.network.neutron [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updated VIF entry in instance network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1127.478653] env[63197]: DEBUG nova.network.neutron [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.574617] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.574873] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.575098] env[63197]: DEBUG nova.objects.instance [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'resources' on Instance uuid ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1127.679124] env[63197]: DEBUG oslo_vmware.api [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364707, 'name': PowerOnVM_Task, 'duration_secs': 0.437748} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.679437] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1127.679650] env[63197]: INFO nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Took 7.22 seconds to spawn the instance on the hypervisor. [ 1127.679831] env[63197]: DEBUG nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1127.680625] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa690ad-a9a6-43af-aea9-4e701d36f8b4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.772544] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 10976c1e-e6ff-4db9-9ef2-a0ff3cc73373] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1127.982146] env[63197]: DEBUG oslo_concurrency.lockutils [req-70ba8cf3-d513-4a6a-a012-bad860459fe0 req-b752ec80-1a6d-4c70-ae01-cefdf5b61a17 service nova] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.078187] env[63197]: DEBUG nova.objects.instance [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'numa_topology' on Instance uuid ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.198910] env[63197]: INFO nova.compute.manager [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Took 11.97 seconds to build instance. [ 1128.276721] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 4672f595-e1f6-4400-b5a1-065598584980] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1128.376947] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc82ecf-b9e0-4a64-92b2-e57a7cf28292 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.386086] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-84be75c5-1d92-4ed3-a670-ec04377fa7b5 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Suspending the VM {{(pid=63197) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1128.386327] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-43134e19-cac0-43eb-bdfe-c558a28c5b8f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.392609] env[63197]: DEBUG oslo_vmware.api [None req-84be75c5-1d92-4ed3-a670-ec04377fa7b5 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1128.392609] env[63197]: value = "task-1364708" [ 1128.392609] env[63197]: _type = "Task" [ 1128.392609] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.400577] env[63197]: DEBUG oslo_vmware.api [None req-84be75c5-1d92-4ed3-a670-ec04377fa7b5 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364708, 'name': SuspendVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.580675] env[63197]: DEBUG nova.objects.base [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1128.645833] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a09d93e-398c-4444-a4cb-d7bdd5461cd6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.654657] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162cc05e-2191-4a3b-b889-44ecddf208da {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.688330] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5892a651-605b-4d43-af32-52d117979c5a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.697467] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de67fbbb-ec28-460e-bf51-9481dbdc9708 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.702096] env[63197]: DEBUG oslo_concurrency.lockutils [None req-641fe098-3a59-48ab-a3fb-4db06225e9a9 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.485s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.712919] env[63197]: DEBUG nova.compute.provider_tree [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1128.780201] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 8e5ef1b0-7532-498a-84c2-189274a36c50] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1128.904209] env[63197]: DEBUG oslo_vmware.api [None req-84be75c5-1d92-4ed3-a670-ec04377fa7b5 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364708, 'name': SuspendVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.217420] env[63197]: DEBUG nova.scheduler.client.report [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1129.283822] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: c41b34cb-cc39-461e-83df-cd4de6780d95] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1129.403932] env[63197]: DEBUG oslo_vmware.api [None req-84be75c5-1d92-4ed3-a670-ec04377fa7b5 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364708, 'name': SuspendVM_Task, 'duration_secs': 0.586491} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.404258] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-84be75c5-1d92-4ed3-a670-ec04377fa7b5 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Suspended the VM {{(pid=63197) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1129.404457] env[63197]: DEBUG nova.compute.manager [None req-84be75c5-1d92-4ed3-a670-ec04377fa7b5 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1129.405244] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d706797b-c78c-4548-ace0-9e7e5a2eaccd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.722791] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.148s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.736183] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.787327] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: c97f59db-b1bf-4744-98ed-b6d7c095207f] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1130.231923] env[63197]: DEBUG oslo_concurrency.lockutils [None req-65281ff5-a1ac-4446-82a1-51a9540e9cfe tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.606s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.232806] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.497s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.232993] env[63197]: INFO nova.compute.manager [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Unshelving [ 1130.290436] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: c6686a09-3711-434b-b69f-a2518366fbed] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1130.777083] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "95e316b6-1162-414f-a6ca-43b37eb6597c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.777349] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.777567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "95e316b6-1162-414f-a6ca-43b37eb6597c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.777760] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.777935] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.779968] env[63197]: INFO nova.compute.manager [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Terminating instance [ 1130.781746] env[63197]: DEBUG nova.compute.manager [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1130.781949] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1130.783065] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2910697d-08a2-4dae-90cc-34fd85dbf102 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.791535] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1130.791826] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d8e6016-9955-444a-8ee1-ba8e43a6044e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.793435] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 4d12bac4-da42-42bd-9361-7015f3be9693] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1130.870634] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1130.870894] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1130.871101] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleting the datastore file [datastore2] 95e316b6-1162-414f-a6ca-43b37eb6597c {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1130.871381] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8d08a81-280d-4b7a-b6fa-ab76ce0eb689 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.879398] env[63197]: DEBUG oslo_vmware.api [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1130.879398] env[63197]: value = "task-1364710" [ 1130.879398] env[63197]: _type = "Task" [ 1130.879398] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.888327] env[63197]: DEBUG oslo_vmware.api [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364710, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.258723] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.258998] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.259231] env[63197]: DEBUG nova.objects.instance [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'pci_requests' on Instance uuid ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.296261] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 4757f91b-21f1-4a16-bda3-729b12ddf86c] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1131.390270] env[63197]: DEBUG oslo_vmware.api [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364710, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149854} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.390512] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1131.390698] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1131.390873] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1131.391061] env[63197]: INFO nova.compute.manager [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1131.391312] env[63197]: DEBUG oslo.service.loopingcall [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.391507] env[63197]: DEBUG nova.compute.manager [-] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1131.391601] env[63197]: DEBUG nova.network.neutron [-] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1131.666452] env[63197]: DEBUG nova.compute.manager [req-8fea5863-b1ab-4bc6-a573-258a8d13cb6f req-2c9a72b8-c778-48ba-8cdc-62a37ead6b9f service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Received event network-vif-deleted-a208670d-3e72-4aef-a6bb-8aba06a94df2 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.666452] env[63197]: INFO nova.compute.manager [req-8fea5863-b1ab-4bc6-a573-258a8d13cb6f req-2c9a72b8-c778-48ba-8cdc-62a37ead6b9f service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Neutron deleted interface a208670d-3e72-4aef-a6bb-8aba06a94df2; detaching it from the instance and deleting it from the info cache [ 1131.666452] env[63197]: DEBUG nova.network.neutron [req-8fea5863-b1ab-4bc6-a573-258a8d13cb6f req-2c9a72b8-c778-48ba-8cdc-62a37ead6b9f service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.762590] env[63197]: DEBUG nova.objects.instance [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'numa_topology' on Instance uuid ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.799198] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 537cc8cc-94e0-41b4-8204-6f995f2ea4d3] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1132.142836] env[63197]: DEBUG nova.network.neutron [-] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.169400] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f37cc683-a515-41c8-b548-4c4ee46c82d2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.181135] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c995aaa-33c8-4d0d-b094-5c29d7c904f1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.207739] env[63197]: DEBUG nova.compute.manager [req-8fea5863-b1ab-4bc6-a573-258a8d13cb6f req-2c9a72b8-c778-48ba-8cdc-62a37ead6b9f service nova] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Detach interface failed, port_id=a208670d-3e72-4aef-a6bb-8aba06a94df2, reason: Instance 95e316b6-1162-414f-a6ca-43b37eb6597c could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1132.265214] env[63197]: INFO nova.compute.claims [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1132.302272] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 52549788-b58f-4f5d-8443-a0c771099884] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1132.645187] env[63197]: INFO nova.compute.manager [-] [instance: 95e316b6-1162-414f-a6ca-43b37eb6597c] Took 1.25 seconds to deallocate network for instance. [ 1132.805045] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: e87c0992-0c7c-4de8-ac68-800eb80cfbc4] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1133.151894] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.308438] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a1f365de-b634-481a-b2e8-9bfc801a47f5] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1133.332750] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5ddacd-5e19-4260-90fd-41be63414225 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.342211] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7359706f-3500-4ee7-9f60-f434f6dc973a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.371813] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4506b21b-767f-4005-ab06-73afd1ebd069 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.379273] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b293b12-dd4c-4f53-8d48-689668be0b72 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.392504] env[63197]: DEBUG nova.compute.provider_tree [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1133.811654] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9d8c131d-1183-4508-ae2d-28e38a50e58d] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1133.895276] env[63197]: DEBUG nova.scheduler.client.report [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.315363] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 180e0da7-f7ee-4fcd-be95-c2bf679278d3] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1134.400237] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.141s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.402367] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.251s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.402604] env[63197]: DEBUG nova.objects.instance [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'resources' on Instance uuid 95e316b6-1162-414f-a6ca-43b37eb6597c {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1134.428988] env[63197]: INFO nova.network.neutron [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1134.818830] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 232ea430-6a11-4dbc-9da9-0f734471d03b] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1134.959100] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5dae20-fbb4-4aa3-ad44-ed38f907b0cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.966750] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbc5857-5796-47c1-bd64-d33691877dce {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.995645] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3de05a-3e94-4b92-818c-7d97df1ea1c5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.002729] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cdb22f-3fbe-420a-9114-4c3595584fe8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.015736] env[63197]: DEBUG nova.compute.provider_tree [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.322493] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 638ef9c9-253b-4958-a660-6c1801408a51] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.519091] env[63197]: DEBUG nova.scheduler.client.report [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1135.817087] env[63197]: DEBUG nova.compute.manager [req-f6591217-3c6a-40fa-ba93-f1ac82ca2e14 req-92896d49-af0f-455a-8652-48c03dba55fc service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-vif-plugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1135.817325] env[63197]: DEBUG oslo_concurrency.lockutils [req-f6591217-3c6a-40fa-ba93-f1ac82ca2e14 req-92896d49-af0f-455a-8652-48c03dba55fc service nova] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.817544] env[63197]: DEBUG oslo_concurrency.lockutils [req-f6591217-3c6a-40fa-ba93-f1ac82ca2e14 req-92896d49-af0f-455a-8652-48c03dba55fc service nova] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.817719] env[63197]: DEBUG oslo_concurrency.lockutils [req-f6591217-3c6a-40fa-ba93-f1ac82ca2e14 req-92896d49-af0f-455a-8652-48c03dba55fc service nova] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.817890] env[63197]: DEBUG nova.compute.manager [req-f6591217-3c6a-40fa-ba93-f1ac82ca2e14 req-92896d49-af0f-455a-8652-48c03dba55fc service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] No waiting events found dispatching network-vif-plugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1135.818069] env[63197]: WARNING nova.compute.manager [req-f6591217-3c6a-40fa-ba93-f1ac82ca2e14 req-92896d49-af0f-455a-8652-48c03dba55fc service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received unexpected event network-vif-plugged-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af for instance with vm_state shelved_offloaded and task_state spawning. [ 1135.825204] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 1067806f-7a4d-4e3f-86c5-7b0d97f33d2d] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.907955] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.908177] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.908343] env[63197]: DEBUG nova.network.neutron [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1136.024292] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.622s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.042655] env[63197]: INFO nova.scheduler.client.report [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted allocations for instance 95e316b6-1162-414f-a6ca-43b37eb6597c [ 1136.328818] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9c80bc51-c8f6-436d-9d19-47f6d1aadbf6] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1136.553029] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e45540f-bd79-47ad-a5a2-19964037cc0d tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "95e316b6-1162-414f-a6ca-43b37eb6597c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.775s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.617795] env[63197]: DEBUG nova.network.neutron [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.832219] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 5b9c6a1c-4212-4941-aa6c-364fb82f5e64] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1137.120858] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1137.143067] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='b8a18f1f2bd960dc4db512e586bf684c',container_format='bare',created_at=2024-10-25T09:52:56Z,direct_url=,disk_format='vmdk',id=967e0b8c-0df0-4299-8693-ceb1aebdc831,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-256528613-shelved',owner='f35fa478269e4de68ae997e79237b25f',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-25T09:53:11Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1137.143333] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1137.143497] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1137.143683] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1137.143832] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1137.143982] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1137.144212] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1137.144422] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1137.144589] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1137.144788] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1137.144975] env[63197]: DEBUG nova.virt.hardware [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1137.145847] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e07df2b-e33e-4ce5-b9fb-921719e29a0b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.154893] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682e462d-90b1-4b73-8575-aee02a9caab4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.168935] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:03:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ec9b6a1b-f520-4f03-b0ce-51ee8346b1af', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1137.176378] env[63197]: DEBUG oslo.service.loopingcall [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1137.176653] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1137.176871] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c00894f5-2459-46ef-a95d-bb0aa637f253 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.197244] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1137.197244] env[63197]: value = "task-1364711" [ 1137.197244] env[63197]: _type = "Task" [ 1137.197244] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.209493] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364711, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.337565] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 8500a7e9-3fdc-411e-a48e-189d4d7bffba] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1137.707969] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364711, 'name': CreateVM_Task, 'duration_secs': 0.311579} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.708164] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1137.708841] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.709028] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.709424] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1137.709683] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6268609-0912-4a2e-a7a5-6f01db06422e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.715378] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.715601] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.718393] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1137.718393] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521f3c3b-2455-a86a-743c-6eaf717f097a" [ 1137.718393] env[63197]: _type = "Task" [ 1137.718393] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.726192] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521f3c3b-2455-a86a-743c-6eaf717f097a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.840958] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 46fd7dec-588e-430e-b51f-9b61d9b148b6] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1137.844888] env[63197]: DEBUG nova.compute.manager [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1137.845091] env[63197]: DEBUG nova.compute.manager [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing instance network info cache due to event network-changed-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1137.845471] env[63197]: DEBUG oslo_concurrency.lockutils [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.845575] env[63197]: DEBUG oslo_concurrency.lockutils [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.845745] env[63197]: DEBUG nova.network.neutron [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Refreshing network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1138.218409] env[63197]: DEBUG nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1138.230728] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.230972] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Processing image 967e0b8c-0df0-4299-8693-ceb1aebdc831 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1138.231225] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831/967e0b8c-0df0-4299-8693-ceb1aebdc831.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.231374] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquired lock "[datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831/967e0b8c-0df0-4299-8693-ceb1aebdc831.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.231556] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1138.231822] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dbea37af-a258-4505-8496-8cb02080df02 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.249880] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1138.250086] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1138.250802] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50332392-4b48-4680-8135-12a41635d070 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.256683] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1138.256683] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d64f3-7a70-f802-192d-a8b9f4a2bf77" [ 1138.256683] env[63197]: _type = "Task" [ 1138.256683] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.264838] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]521d64f3-7a70-f802-192d-a8b9f4a2bf77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.347514] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: a0a40337-d8d7-448b-afff-f6849e9d37a1] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.558782] env[63197]: DEBUG nova.network.neutron [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updated VIF entry in instance network info cache for port ec9b6a1b-f520-4f03-b0ce-51ee8346b1af. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1138.559169] env[63197]: DEBUG nova.network.neutron [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.744095] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.744382] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.746507] env[63197]: INFO nova.compute.claims [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1138.768067] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Preparing fetch location {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1138.768067] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Fetch image to [datastore2] OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb/OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb.vmdk {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1138.768067] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Downloading stream optimized image 967e0b8c-0df0-4299-8693-ceb1aebdc831 to [datastore2] OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb/OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb.vmdk on the data store datastore2 as vApp {{(pid=63197) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1138.768291] env[63197]: DEBUG nova.virt.vmwareapi.images [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Downloading image file data 967e0b8c-0df0-4299-8693-ceb1aebdc831 to the ESX as VM named 'OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb' {{(pid=63197) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1138.838357] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1138.838357] env[63197]: value = "resgroup-9" [ 1138.838357] env[63197]: _type = "ResourcePool" [ 1138.838357] env[63197]: }. {{(pid=63197) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1138.838662] env[63197]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-09c26b8c-4de5-43e7-9673-da2d1049d003 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.853679] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 72f9b0c4-69b9-49f2-8665-ff77151883af] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.862240] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease: (returnval){ [ 1138.862240] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c045c5-ec05-4af2-259d-1e58613f3dea" [ 1138.862240] env[63197]: _type = "HttpNfcLease" [ 1138.862240] env[63197]: } obtained for vApp import into resource pool (val){ [ 1138.862240] env[63197]: value = "resgroup-9" [ 1138.862240] env[63197]: _type = "ResourcePool" [ 1138.862240] env[63197]: }. {{(pid=63197) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1138.862502] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the lease: (returnval){ [ 1138.862502] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c045c5-ec05-4af2-259d-1e58613f3dea" [ 1138.862502] env[63197]: _type = "HttpNfcLease" [ 1138.862502] env[63197]: } to be ready. {{(pid=63197) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1138.869279] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1138.869279] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c045c5-ec05-4af2-259d-1e58613f3dea" [ 1138.869279] env[63197]: _type = "HttpNfcLease" [ 1138.869279] env[63197]: } is initializing. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1139.061826] env[63197]: DEBUG oslo_concurrency.lockutils [req-9b6f598e-b8b3-4757-893b-e409621dd059 req-def6f57b-0951-4d41-b680-bc59e6dda6ab service nova] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.357197] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 9c1caa2b-c369-425a-8726-cddadf06f338] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.371488] env[63197]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1139.371488] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c045c5-ec05-4af2-259d-1e58613f3dea" [ 1139.371488] env[63197]: _type = "HttpNfcLease" [ 1139.371488] env[63197]: } is ready. {{(pid=63197) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1139.372029] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1139.372029] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52c045c5-ec05-4af2-259d-1e58613f3dea" [ 1139.372029] env[63197]: _type = "HttpNfcLease" [ 1139.372029] env[63197]: }. {{(pid=63197) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1139.373025] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26482379-98e9-4f3e-b35a-ad147178a1e9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.380161] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a0d29b-9e6f-5f5a-1d24-75872431acee/disk-0.vmdk from lease info. {{(pid=63197) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1139.380350] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a0d29b-9e6f-5f5a-1d24-75872431acee/disk-0.vmdk. {{(pid=63197) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1139.444081] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ef667ab9-0f11-4d80-b235-856505cc7e77 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.825370] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1502eef-a90a-4c34-a39b-ab620f1347f9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.837042] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad039f5-8d79-4b38-9f33-55ccf543d71a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.871397] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 3b7aa0b0-d115-4026-a984-0e9fc73f2d2e] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.878817] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f565a1e-c52b-4f7a-b0ce-57453ff66b14 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.889825] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2819e489-833b-459c-8dc9-a8b46fb2e10f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.904790] env[63197]: DEBUG nova.compute.provider_tree [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.384241] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: c55b5b2b-5ecc-43bb-a279-7370cd9ac722] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.408753] env[63197]: DEBUG nova.scheduler.client.report [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1140.456295] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Completed reading data from the image iterator. {{(pid=63197) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1140.456590] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a0d29b-9e6f-5f5a-1d24-75872431acee/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1140.457649] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e229091-3d49-43a2-b29f-8846fb55d56d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.465408] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a0d29b-9e6f-5f5a-1d24-75872431acee/disk-0.vmdk is in state: ready. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1140.465607] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a0d29b-9e6f-5f5a-1d24-75872431acee/disk-0.vmdk. {{(pid=63197) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1140.465867] env[63197]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4627bf82-5ced-44fa-8cd5-7622aaa8a788 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.668096] env[63197]: DEBUG oslo_vmware.rw_handles [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a0d29b-9e6f-5f5a-1d24-75872431acee/disk-0.vmdk. {{(pid=63197) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1140.668324] env[63197]: INFO nova.virt.vmwareapi.images [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Downloaded image file data 967e0b8c-0df0-4299-8693-ceb1aebdc831 [ 1140.669156] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e667d6e-7c00-4e57-8c6e-82582b095a2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.683985] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21cc1ad6-6946-44cc-8ae0-3d89a65625ac {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.720157] env[63197]: INFO nova.virt.vmwareapi.images [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] The imported VM was unregistered [ 1140.722555] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Caching image {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1140.722797] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Creating directory with path [datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1140.723065] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea8aaac7-5870-4e66-86bd-ac1566ac17a5 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.742072] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Created directory with path [datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831 {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1140.742254] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb/OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb.vmdk to [datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831/967e0b8c-0df0-4299-8693-ceb1aebdc831.vmdk. {{(pid=63197) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1140.742514] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-1ebd5f8d-35a2-4b47-813c-47e4cb28ccb4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.748959] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1140.748959] env[63197]: value = "task-1364714" [ 1140.748959] env[63197]: _type = "Task" [ 1140.748959] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.758444] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364714, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.888717] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 65df32fa-a6e7-4c3a-af8b-0422df8d9229] Instance has had 0 of 5 cleanup attempts {{(pid=63197) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.915355] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.171s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.915961] env[63197]: DEBUG nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1141.259443] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364714, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.393726] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.393726] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Cleaning up deleted instances with incomplete migration {{(pid=63197) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1141.421103] env[63197]: DEBUG nova.compute.utils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1141.423160] env[63197]: DEBUG nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1141.423398] env[63197]: DEBUG nova.network.neutron [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1141.472723] env[63197]: DEBUG nova.policy [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57a8e87b64fe46d7ab7e570d57611119', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0ceb1016d6d34bff8880dca42d495377', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1141.759837] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364714, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.895754] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.923911] env[63197]: DEBUG nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1141.986063] env[63197]: DEBUG nova.network.neutron [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Successfully created port: 3cecd3aa-c700-408e-b8ce-50c12cc106a0 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1142.263132] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364714, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.761185] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364714, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.933936] env[63197]: DEBUG nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1142.961407] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1142.961665] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1142.961833] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1142.962371] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1142.962371] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1142.962371] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1142.962524] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1142.962762] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1142.962950] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1142.963137] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1142.963319] env[63197]: DEBUG nova.virt.hardware [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1142.964263] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a04fbe3-80a6-45a3-a8da-1010ad89d084 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.972183] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a751a505-0b71-42d9-95ec-9189bac5f8c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.261117] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364714, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.274209} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.261392] env[63197]: INFO nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb/OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb.vmdk to [datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831/967e0b8c-0df0-4299-8693-ceb1aebdc831.vmdk. [ 1143.261583] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Cleaning up location [datastore2] OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1143.261750] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_ea5de86a-6763-49e7-a2ff-13fe2d662aeb {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.262023] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b402a54-5f07-4646-b468-3ef1db927e95 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.268209] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1143.268209] env[63197]: value = "task-1364715" [ 1143.268209] env[63197]: _type = "Task" [ 1143.268209] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.275194] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364715, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.437643] env[63197]: DEBUG nova.compute.manager [req-31a4ce24-b4ca-4319-9850-ac4a75a5e10d req-3c336c40-03e8-4d0f-be98-54ca73a639ff service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Received event network-vif-plugged-3cecd3aa-c700-408e-b8ce-50c12cc106a0 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1143.437915] env[63197]: DEBUG oslo_concurrency.lockutils [req-31a4ce24-b4ca-4319-9850-ac4a75a5e10d req-3c336c40-03e8-4d0f-be98-54ca73a639ff service nova] Acquiring lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.438150] env[63197]: DEBUG oslo_concurrency.lockutils [req-31a4ce24-b4ca-4319-9850-ac4a75a5e10d req-3c336c40-03e8-4d0f-be98-54ca73a639ff service nova] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.438324] env[63197]: DEBUG oslo_concurrency.lockutils [req-31a4ce24-b4ca-4319-9850-ac4a75a5e10d req-3c336c40-03e8-4d0f-be98-54ca73a639ff service nova] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.438490] env[63197]: DEBUG nova.compute.manager [req-31a4ce24-b4ca-4319-9850-ac4a75a5e10d req-3c336c40-03e8-4d0f-be98-54ca73a639ff service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] No waiting events found dispatching network-vif-plugged-3cecd3aa-c700-408e-b8ce-50c12cc106a0 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1143.438740] env[63197]: WARNING nova.compute.manager [req-31a4ce24-b4ca-4319-9850-ac4a75a5e10d req-3c336c40-03e8-4d0f-be98-54ca73a639ff service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Received unexpected event network-vif-plugged-3cecd3aa-c700-408e-b8ce-50c12cc106a0 for instance with vm_state building and task_state spawning. [ 1143.522842] env[63197]: DEBUG nova.network.neutron [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Successfully updated port: 3cecd3aa-c700-408e-b8ce-50c12cc106a0 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1143.777782] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364715, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031024} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.778050] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.778254] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Releasing lock "[datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831/967e0b8c-0df0-4299-8693-ceb1aebdc831.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.778518] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831/967e0b8c-0df0-4299-8693-ceb1aebdc831.vmdk to [datastore2] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1143.778772] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0963a8c-e4de-4e50-a694-acca0020f0dc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.785213] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1143.785213] env[63197]: value = "task-1364716" [ 1143.785213] env[63197]: _type = "Task" [ 1143.785213] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.793157] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364716, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.890090] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.890295] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1143.890389] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Rebuilding the list of instances to heal {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1144.025813] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.026159] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.026307] env[63197]: DEBUG nova.network.neutron [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1144.295813] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364716, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.395044] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Skipping network cache update for instance because it is Building. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1144.395044] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.395247] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquired lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.395307] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Forcefully refreshing network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1144.395436] env[63197]: DEBUG nova.objects.instance [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lazy-loading 'info_cache' on Instance uuid ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.565990] env[63197]: DEBUG nova.network.neutron [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1144.739885] env[63197]: DEBUG nova.network.neutron [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance_info_cache with network_info: [{"id": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "address": "fa:16:3e:b4:f5:5e", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cecd3aa-c7", "ovs_interfaceid": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.797095] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364716, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.242977] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.243385] env[63197]: DEBUG nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Instance network_info: |[{"id": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "address": "fa:16:3e:b4:f5:5e", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cecd3aa-c7", "ovs_interfaceid": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1145.243842] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:f5:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3cecd3aa-c700-408e-b8ce-50c12cc106a0', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1145.251575] env[63197]: DEBUG oslo.service.loopingcall [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1145.251864] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1145.252127] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7170be4-751c-4e2d-9be6-ad1125cdd749 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.272969] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1145.272969] env[63197]: value = "task-1364717" [ 1145.272969] env[63197]: _type = "Task" [ 1145.272969] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.281056] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364717, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.296026] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364716, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.464997] env[63197]: DEBUG nova.compute.manager [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Received event network-changed-3cecd3aa-c700-408e-b8ce-50c12cc106a0 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1145.464997] env[63197]: DEBUG nova.compute.manager [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Refreshing instance network info cache due to event network-changed-3cecd3aa-c700-408e-b8ce-50c12cc106a0. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1145.465418] env[63197]: DEBUG oslo_concurrency.lockutils [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] Acquiring lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.465418] env[63197]: DEBUG oslo_concurrency.lockutils [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] Acquired lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.465526] env[63197]: DEBUG nova.network.neutron [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Refreshing network info cache for port 3cecd3aa-c700-408e-b8ce-50c12cc106a0 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1145.783275] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364717, 'name': CreateVM_Task} progress is 25%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.797302] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364716, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.220804] env[63197]: DEBUG nova.network.neutron [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [{"id": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "address": "fa:16:3e:0e:03:28", "network": {"id": "52cb0057-5f4c-4de2-9e8d-cac718b6ca47", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-550451236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f35fa478269e4de68ae997e79237b25f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapec9b6a1b-f5", "ovs_interfaceid": "ec9b6a1b-f520-4f03-b0ce-51ee8346b1af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.257175] env[63197]: DEBUG nova.network.neutron [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updated VIF entry in instance network info cache for port 3cecd3aa-c700-408e-b8ce-50c12cc106a0. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1146.257557] env[63197]: DEBUG nova.network.neutron [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance_info_cache with network_info: [{"id": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "address": "fa:16:3e:b4:f5:5e", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cecd3aa-c7", "ovs_interfaceid": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.283467] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364717, 'name': CreateVM_Task, 'duration_secs': 0.846139} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.283622] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1146.284256] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.284428] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.284769] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1146.285028] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaf596a8-a0a0-4e23-8c2c-59644f8433ad {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.291760] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1146.291760] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ce0115-c241-5ba8-a441-9c6717dbc7bf" [ 1146.291760] env[63197]: _type = "Task" [ 1146.291760] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.297379] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364716, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.203229} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.297892] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/967e0b8c-0df0-4299-8693-ceb1aebdc831/967e0b8c-0df0-4299-8693-ceb1aebdc831.vmdk to [datastore2] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1146.298651] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8956f053-ff7b-46cd-932d-ea7a5cc986d9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.304213] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52ce0115-c241-5ba8-a441-9c6717dbc7bf, 'name': SearchDatastore_Task, 'duration_secs': 0.009032} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.304777] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.305010] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1146.305238] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.305388] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.305569] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1146.305796] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5382c203-968b-4824-a5c8-ee6e3aa283f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.323940] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk or device None with type streamOptimized {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1146.324491] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a13ac5c3-a019-4b70-b165-8e89d040b7a7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.339767] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1146.339951] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1146.340929] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b822f28-cb04-4340-8e85-637b81c972c7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.345578] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1146.345578] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522f6bfc-917f-0c91-0067-20b5fb4ca8dc" [ 1146.345578] env[63197]: _type = "Task" [ 1146.345578] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.346732] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1146.346732] env[63197]: value = "task-1364718" [ 1146.346732] env[63197]: _type = "Task" [ 1146.346732] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.359523] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]522f6bfc-917f-0c91-0067-20b5fb4ca8dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008513} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.359730] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364718, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.360444] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b899ab4-b6fb-451b-9dac-e185434f22b9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.364867] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1146.364867] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a781bc-6a1d-997c-8b4b-15e06d6f651a" [ 1146.364867] env[63197]: _type = "Task" [ 1146.364867] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.372288] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a781bc-6a1d-997c-8b4b-15e06d6f651a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.723620] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Releasing lock "refresh_cache-ad216370-3f6d-4ebe-a956-7dd3bf05511f" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.723854] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updated the network info_cache for instance {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1146.724132] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.760175] env[63197]: DEBUG oslo_concurrency.lockutils [req-83fd4451-45a6-4542-b630-70bdecc77593 req-8cce9bdf-0671-4b35-b9ba-9953001aff1b service nova] Releasing lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.859571] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364718, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.873474] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52a781bc-6a1d-997c-8b4b-15e06d6f651a, 'name': SearchDatastore_Task, 'duration_secs': 0.00837} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.873732] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.873996] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] fae5ada6-f3ed-4428-8ea8-979ca98d2e0e/fae5ada6-f3ed-4428-8ea8-979ca98d2e0e.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1146.874271] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d20390e4-70cf-4b7f-b24c-651daa49f6b1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.882471] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1146.882471] env[63197]: value = "task-1364719" [ 1146.882471] env[63197]: _type = "Task" [ 1146.882471] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.890329] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364719, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.359423] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364718, 'name': ReconfigVM_Task, 'duration_secs': 0.671597} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.359829] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Reconfigured VM instance instance-00000063 to attach disk [datastore2] ad216370-3f6d-4ebe-a956-7dd3bf05511f/ad216370-3f6d-4ebe-a956-7dd3bf05511f.vmdk or device None with type streamOptimized {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1147.360299] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ecbcdd8-4810-4333-9de2-9027dfc21683 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.366371] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1147.366371] env[63197]: value = "task-1364720" [ 1147.366371] env[63197]: _type = "Task" [ 1147.366371] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.374104] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364720, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.393208] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364719, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450915} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.393535] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] fae5ada6-f3ed-4428-8ea8-979ca98d2e0e/fae5ada6-f3ed-4428-8ea8-979ca98d2e0e.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1147.393763] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1147.394049] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b02db21-fb11-4596-a36d-af2ec850169e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.399609] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1147.399609] env[63197]: value = "task-1364721" [ 1147.399609] env[63197]: _type = "Task" [ 1147.399609] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.406769] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364721, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.876242] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364720, 'name': Rename_Task, 'duration_secs': 0.132489} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.876467] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1147.876710] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0c27241-0314-4035-a022-2c31f9441b5c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.882795] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1147.882795] env[63197]: value = "task-1364722" [ 1147.882795] env[63197]: _type = "Task" [ 1147.882795] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.889740] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364722, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.906558] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364721, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05991} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.906792] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1147.907564] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a763e2-f3dc-4747-a423-2e4b3454f91f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.928547] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] fae5ada6-f3ed-4428-8ea8-979ca98d2e0e/fae5ada6-f3ed-4428-8ea8-979ca98d2e0e.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.928810] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c40ca4f-cd47-4a1a-870f-7626d201ac3c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.946858] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1147.946858] env[63197]: value = "task-1364723" [ 1147.946858] env[63197]: _type = "Task" [ 1147.946858] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.955735] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364723, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.393414] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364722, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.456678] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364723, 'name': ReconfigVM_Task, 'duration_secs': 0.268817} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.457100] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Reconfigured VM instance instance-00000069 to attach disk [datastore2] fae5ada6-f3ed-4428-8ea8-979ca98d2e0e/fae5ada6-f3ed-4428-8ea8-979ca98d2e0e.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.457776] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e93a3de-1b8c-4364-923d-74f2660ce08c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.463959] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1148.463959] env[63197]: value = "task-1364724" [ 1148.463959] env[63197]: _type = "Task" [ 1148.463959] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.471840] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364724, 'name': Rename_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.894220] env[63197]: DEBUG oslo_vmware.api [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364722, 'name': PowerOnVM_Task, 'duration_secs': 0.562616} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.894484] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1148.973433] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364724, 'name': Rename_Task, 'duration_secs': 0.140113} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.973752] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1148.973968] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea37a2ac-924d-48cc-b51a-17d5d7639b2a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.980054] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1148.980054] env[63197]: value = "task-1364725" [ 1148.980054] env[63197]: _type = "Task" [ 1148.980054] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.988366] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.989217] env[63197]: DEBUG nova.compute.manager [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1148.989923] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c83d51a-f321-463e-a8b8-08e53e2e1356 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.490249] env[63197]: DEBUG oslo_vmware.api [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364725, 'name': PowerOnVM_Task, 'duration_secs': 0.443238} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.490634] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1149.490709] env[63197]: INFO nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Took 6.56 seconds to spawn the instance on the hypervisor. [ 1149.490890] env[63197]: DEBUG nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.491672] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b573d2e8-0fb5-4d86-8d7e-cd1c558dfcfd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.506798] env[63197]: DEBUG oslo_concurrency.lockutils [None req-dc977229-05d7-4430-84e1-65af71883dc0 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.274s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.009889] env[63197]: INFO nova.compute.manager [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Took 11.28 seconds to build instance. [ 1150.512084] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e762a7e-68f9-43ea-81c3-007ea89d020f tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.796s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.588959] env[63197]: DEBUG nova.compute.manager [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Stashing vm_state: active {{(pid=63197) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1152.108696] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.108963] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.615013] env[63197]: INFO nova.compute.claims [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1153.122209] env[63197]: INFO nova.compute.resource_tracker [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating resource usage from migration b65c0df8-5de5-42d3-80b1-146ced5d985e [ 1153.190733] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e376d506-b0d2-471b-a2a2-b9a957b4e09f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.198254] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51330957-a061-47f3-b2e1-776f65b23413 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.228698] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79229149-8260-44ca-879e-4007165477d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.235765] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c01a494-c41f-4716-b870-27a1e5655966 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.248450] env[63197]: DEBUG nova.compute.provider_tree [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.751390] env[63197]: DEBUG nova.scheduler.client.report [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1154.258271] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.149s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.258535] env[63197]: INFO nova.compute.manager [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Migrating [ 1154.775276] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1154.775678] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.775678] env[63197]: DEBUG nova.network.neutron [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1155.466648] env[63197]: DEBUG nova.network.neutron [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance_info_cache with network_info: [{"id": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "address": "fa:16:3e:b4:f5:5e", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cecd3aa-c7", "ovs_interfaceid": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.969765] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.332725] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.332967] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.836329] env[63197]: DEBUG nova.compute.utils [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1157.339205] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.485638] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e568266-33ff-40c6-84e4-ac1f0d5cca0e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.504848] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance 'fae5ada6-f3ed-4428-8ea8-979ca98d2e0e' progress to 0 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1158.010755] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1158.011095] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cd1e747-7f48-482d-a9de-2773d27e43e1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.019943] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1158.019943] env[63197]: value = "task-1364726" [ 1158.019943] env[63197]: _type = "Task" [ 1158.019943] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.029397] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364726, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.403128] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.403477] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.403611] env[63197]: INFO nova.compute.manager [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Attaching volume 11297d4a-4378-4839-9470-e6876cf68883 to /dev/sdb [ 1158.435443] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a02186-c6b2-47fc-9879-b697b051b41e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.443063] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0934ae8f-524d-43c3-91e1-feb3beb00278 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.458023] env[63197]: DEBUG nova.virt.block_device [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updating existing volume attachment record: 84486625-2384-4ad4-9aa7-3ee647df21bc {{(pid=63197) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1158.528880] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364726, 'name': PowerOffVM_Task, 'duration_secs': 0.19704} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.529181] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1158.529376] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance 'fae5ada6-f3ed-4428-8ea8-979ca98d2e0e' progress to 17 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1159.035792] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:11Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1159.036078] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1159.036226] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1159.036421] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1159.036576] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1159.036729] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1159.036991] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1159.037176] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1159.037378] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1159.037556] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1159.037734] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1159.042746] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3368bdc6-4177-4129-9713-7c475475eb8d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.058463] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1159.058463] env[63197]: value = "task-1364728" [ 1159.058463] env[63197]: _type = "Task" [ 1159.058463] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.066799] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364728, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.568640] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364728, 'name': ReconfigVM_Task, 'duration_secs': 0.134844} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.569036] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance 'fae5ada6-f3ed-4428-8ea8-979ca98d2e0e' progress to 33 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1160.004572] env[63197]: INFO nova.compute.manager [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Rebuilding instance [ 1160.042651] env[63197]: DEBUG nova.compute.manager [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1160.043555] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4015ab1-9d3b-4573-a35d-f092219665cd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.075444] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1160.075741] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1160.075922] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1160.076127] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1160.076281] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1160.076447] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1160.076668] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1160.076837] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1160.077014] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1160.077209] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1160.077391] env[63197]: DEBUG nova.virt.hardware [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1160.082661] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1160.082927] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b693ae66-137c-4d12-a055-8afabb1f7e02 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.102184] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1160.102184] env[63197]: value = "task-1364729" [ 1160.102184] env[63197]: _type = "Task" [ 1160.102184] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.110398] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364729, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.554189] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1160.554522] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6543e1f2-9349-4287-abfc-788949fa417c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.561550] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1160.561550] env[63197]: value = "task-1364730" [ 1160.561550] env[63197]: _type = "Task" [ 1160.561550] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.569397] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364730, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.611275] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364729, 'name': ReconfigVM_Task, 'duration_secs': 0.146342} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.611508] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1160.612652] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9eaa2f-77b6-455e-9dc4-bce05da90c67 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.636095] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] fae5ada6-f3ed-4428-8ea8-979ca98d2e0e/fae5ada6-f3ed-4428-8ea8-979ca98d2e0e.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1160.636367] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e3d790b8-8bc0-49f6-a36e-b33ba9432e2c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.654285] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1160.654285] env[63197]: value = "task-1364731" [ 1160.654285] env[63197]: _type = "Task" [ 1160.654285] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.662172] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364731, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.071875] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364730, 'name': PowerOffVM_Task, 'duration_secs': 0.159059} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.072175] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1161.072851] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.073120] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-365e4c3f-0f0f-466d-95a4-837e3a6531ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.079070] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1161.079070] env[63197]: value = "task-1364733" [ 1161.079070] env[63197]: _type = "Task" [ 1161.079070] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.086134] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364733, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.163124] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364731, 'name': ReconfigVM_Task, 'duration_secs': 0.252594} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.163417] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Reconfigured VM instance instance-00000069 to attach disk [datastore2] fae5ada6-f3ed-4428-8ea8-979ca98d2e0e/fae5ada6-f3ed-4428-8ea8-979ca98d2e0e.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1161.163688] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance 'fae5ada6-f3ed-4428-8ea8-979ca98d2e0e' progress to 50 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1161.589991] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] VM already powered off {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1161.590598] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1161.590598] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290421', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'name': 'volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '14456558-a814-45cd-9596-0c073c581b57', 'attached_at': '', 'detached_at': '', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'serial': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1161.591269] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226ec9f3-fba3-424a-ac43-4261e17048f2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.608371] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c77952-f25e-4291-a28f-bde250b1d75b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.614766] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b2551f-4c8b-4d08-b82b-768a8c05e5eb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.632030] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b218284f-db31-490a-ac37-718aed9d8bba {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.646256] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] The volume has not been displaced from its original location: [datastore2] volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a/volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1161.651396] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1161.651690] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5cf32b1-cc9f-4934-b437-4653cc601a68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.670063] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1161.670063] env[63197]: value = "task-1364734" [ 1161.670063] env[63197]: _type = "Task" [ 1161.670063] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.670756] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8d725a-b930-4b4e-940c-fda7a10e8e2f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.693881] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364734, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.694750] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f640ffc7-9cae-4fae-9d0f-862fda703024 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.712783] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance 'fae5ada6-f3ed-4428-8ea8-979ca98d2e0e' progress to 67 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1162.183022] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364734, 'name': ReconfigVM_Task, 'duration_secs': 0.335159} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.183315] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1162.187995] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35cff4dc-ce45-491d-a776-2e4ec3368be8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.202826] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1162.202826] env[63197]: value = "task-1364735" [ 1162.202826] env[63197]: _type = "Task" [ 1162.202826] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.210204] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364735, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.253150] env[63197]: DEBUG nova.network.neutron [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Port 3cecd3aa-c700-408e-b8ce-50c12cc106a0 binding to destination host cpu-1 is already ACTIVE {{(pid=63197) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1162.713158] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364735, 'name': ReconfigVM_Task, 'duration_secs': 0.105075} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.713509] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290421', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'name': 'volume-5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '14456558-a814-45cd-9596-0c073c581b57', 'attached_at': '', 'detached_at': '', 'volume_id': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a', 'serial': '5960526b-c5d9-4c8b-ad0b-ba59179fa82a'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1162.713903] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1162.714754] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8c62b1-3090-44a1-9986-4e003acccc1f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.721682] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1162.722072] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-992127c7-9543-4649-960e-0e4bb3e57ea2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.788526] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1162.788760] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1162.788937] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Deleting the datastore file [datastore2] 14456558-a814-45cd-9596-0c073c581b57 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1162.789231] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d425bc83-e815-49f2-a19f-6d6e780e358b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.795816] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for the task: (returnval){ [ 1162.795816] env[63197]: value = "task-1364737" [ 1162.795816] env[63197]: _type = "Task" [ 1162.795816] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.804071] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.002503] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Volume attach. Driver type: vmdk {{(pid=63197) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1163.002809] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290432', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'name': 'volume-11297d4a-4378-4839-9470-e6876cf68883', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0dc57420-b13f-4305-ae1f-6ebf61f356f3', 'attached_at': '', 'detached_at': '', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'serial': '11297d4a-4378-4839-9470-e6876cf68883'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1163.003728] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d41d9c-5483-4180-9692-f797011966dd {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.022042] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a93d4e-f86e-45e7-8335-9beaaf3ef58a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.046036] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-11297d4a-4378-4839-9470-e6876cf68883/volume-11297d4a-4378-4839-9470-e6876cf68883.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1163.046279] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd294a1d-f350-4bb5-a02e-12e932f52cfb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.063744] env[63197]: DEBUG oslo_vmware.api [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1163.063744] env[63197]: value = "task-1364738" [ 1163.063744] env[63197]: _type = "Task" [ 1163.063744] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.071302] env[63197]: DEBUG oslo_vmware.api [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364738, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.276894] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.277307] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.277603] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.305173] env[63197]: DEBUG oslo_vmware.api [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Task: {'id': task-1364737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.065768} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.305417] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1163.305614] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1163.305857] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1163.378170] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1163.378515] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5fa0bd0-cf35-4c67-a558-3f6ea9691c27 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.387020] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece4d44f-6051-465a-b39e-7378adc50798 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.412101] env[63197]: ERROR nova.compute.manager [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Failed to detach volume 5960526b-c5d9-4c8b-ad0b-ba59179fa82a from /dev/sda: nova.exception.InstanceNotFound: Instance 14456558-a814-45cd-9596-0c073c581b57 could not be found. [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] Traceback (most recent call last): [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self.driver.rebuild(**kwargs) [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] raise NotImplementedError() [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] NotImplementedError [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] During handling of the above exception, another exception occurred: [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] Traceback (most recent call last): [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self.driver.detach_volume(context, old_connection_info, [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] return self._volumeops.detach_volume(connection_info, instance) [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self._detach_volume_vmdk(connection_info, instance) [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] stable_ref.fetch_moref(session) [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] nova.exception.InstanceNotFound: Instance 14456558-a814-45cd-9596-0c073c581b57 could not be found. [ 1163.412101] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.531879] env[63197]: DEBUG nova.compute.utils [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Build of instance 14456558-a814-45cd-9596-0c073c581b57 aborted: Failed to rebuild volume backed instance. {{(pid=63197) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1163.534208] env[63197]: ERROR nova.compute.manager [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 14456558-a814-45cd-9596-0c073c581b57 aborted: Failed to rebuild volume backed instance. [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] Traceback (most recent call last): [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self.driver.rebuild(**kwargs) [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] raise NotImplementedError() [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] NotImplementedError [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] During handling of the above exception, another exception occurred: [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] Traceback (most recent call last): [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self._detach_root_volume(context, instance, root_bdm) [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] with excutils.save_and_reraise_exception(): [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self.force_reraise() [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] raise self.value [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self.driver.detach_volume(context, old_connection_info, [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] return self._volumeops.detach_volume(connection_info, instance) [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self._detach_volume_vmdk(connection_info, instance) [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] stable_ref.fetch_moref(session) [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] nova.exception.InstanceNotFound: Instance 14456558-a814-45cd-9596-0c073c581b57 could not be found. [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] During handling of the above exception, another exception occurred: [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] Traceback (most recent call last): [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] yield [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1163.534208] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self._do_rebuild_instance_with_claim( [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self._do_rebuild_instance( [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self._rebuild_default_impl(**kwargs) [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] self._rebuild_volume_backed_instance( [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] raise exception.BuildAbortException( [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] nova.exception.BuildAbortException: Build of instance 14456558-a814-45cd-9596-0c073c581b57 aborted: Failed to rebuild volume backed instance. [ 1163.535457] env[63197]: ERROR nova.compute.manager [instance: 14456558-a814-45cd-9596-0c073c581b57] [ 1163.574012] env[63197]: DEBUG oslo_vmware.api [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364738, 'name': ReconfigVM_Task, 'duration_secs': 0.333982} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.574302] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-11297d4a-4378-4839-9470-e6876cf68883/volume-11297d4a-4378-4839-9470-e6876cf68883.vmdk or device None with type thin {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.579102] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e0e1c42-45b2-48ed-9190-decef767360f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.594908] env[63197]: DEBUG oslo_vmware.api [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1163.594908] env[63197]: value = "task-1364739" [ 1163.594908] env[63197]: _type = "Task" [ 1163.594908] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.604093] env[63197]: DEBUG oslo_vmware.api [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364739, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.107630] env[63197]: DEBUG oslo_vmware.api [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364739, 'name': ReconfigVM_Task, 'duration_secs': 0.129688} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.107974] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290432', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'name': 'volume-11297d4a-4378-4839-9470-e6876cf68883', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0dc57420-b13f-4305-ae1f-6ebf61f356f3', 'attached_at': '', 'detached_at': '', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'serial': '11297d4a-4378-4839-9470-e6876cf68883'} {{(pid=63197) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1164.314464] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.314659] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.314880] env[63197]: DEBUG nova.network.neutron [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1165.037322] env[63197]: DEBUG nova.network.neutron [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance_info_cache with network_info: [{"id": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "address": "fa:16:3e:b4:f5:5e", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cecd3aa-c7", "ovs_interfaceid": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.141220] env[63197]: DEBUG nova.objects.instance [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'flavor' on Instance uuid 0dc57420-b13f-4305-ae1f-6ebf61f356f3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1165.539876] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.548178] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.548435] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.608549] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-500b54f3-caaf-4e5b-9e93-020336490341 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.616295] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b75f5f1-e297-41e4-8379-b668489b26d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.647680] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c616fc-6224-4ae7-9e45-5d1c4ba84499 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.650263] env[63197]: DEBUG oslo_concurrency.lockutils [None req-8cab26af-f461-45b9-bf86-e3d91a8ae9c6 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.655843] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c6d815-93ca-4434-86f3-a1e7dc02c2fc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.669019] env[63197]: DEBUG nova.compute.provider_tree [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.819397] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.819682] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.828546] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquiring lock "14456558-a814-45cd-9596-0c073c581b57" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.828793] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "14456558-a814-45cd-9596-0c073c581b57" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.828970] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquiring lock "14456558-a814-45cd-9596-0c073c581b57-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.829174] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "14456558-a814-45cd-9596-0c073c581b57-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.829345] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "14456558-a814-45cd-9596-0c073c581b57-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.831471] env[63197]: INFO nova.compute.manager [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Terminating instance [ 1165.833316] env[63197]: DEBUG nova.compute.manager [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1165.833827] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-310a6408-a611-46ae-9dbb-c773228bda0d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.843611] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3934711-be8f-433a-9c67-da81b367146f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.867968] env[63197]: WARNING nova.virt.vmwareapi.driver [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 14456558-a814-45cd-9596-0c073c581b57 could not be found. [ 1165.868197] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1165.868490] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3dbedf9-2316-44e0-86d4-867ae4dfbd61 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.876476] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2825787f-b724-4fad-a13b-106265ae3d7b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.899689] env[63197]: WARNING nova.virt.vmwareapi.vmops [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 14456558-a814-45cd-9596-0c073c581b57 could not be found. [ 1165.899845] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1165.900036] env[63197]: INFO nova.compute.manager [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Took 0.07 seconds to destroy the instance on the hypervisor. [ 1165.900279] env[63197]: DEBUG oslo.service.loopingcall [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1165.900498] env[63197]: DEBUG nova.compute.manager [-] [instance: 14456558-a814-45cd-9596-0c073c581b57] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1165.900586] env[63197]: DEBUG nova.network.neutron [-] [instance: 14456558-a814-45cd-9596-0c073c581b57] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1166.060525] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aaee18f-906a-411f-b242-059bfa23b316 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.081533] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1108dfff-de58-42d3-9281-099ebb7158a4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.089135] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance 'fae5ada6-f3ed-4428-8ea8-979ca98d2e0e' progress to 83 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1166.172125] env[63197]: DEBUG nova.scheduler.client.report [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1166.322721] env[63197]: INFO nova.compute.manager [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Detaching volume 11297d4a-4378-4839-9470-e6876cf68883 [ 1166.357716] env[63197]: INFO nova.virt.block_device [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Attempting to driver detach volume 11297d4a-4378-4839-9470-e6876cf68883 from mountpoint /dev/sdb [ 1166.357965] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Volume detach. Driver type: vmdk {{(pid=63197) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1166.358167] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290432', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'name': 'volume-11297d4a-4378-4839-9470-e6876cf68883', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0dc57420-b13f-4305-ae1f-6ebf61f356f3', 'attached_at': '', 'detached_at': '', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'serial': '11297d4a-4378-4839-9470-e6876cf68883'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1166.359190] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809d8dbd-059b-4978-80d6-e3baa3a7e326 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.387280] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d546cca0-2c64-4bef-8642-86a0c4d1e72c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.394737] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d06917-bf25-4495-b7ff-b05ff90b78fb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.415820] env[63197]: DEBUG nova.network.neutron [-] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.417436] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5201066-e9d5-452d-b705-0d8d22a50913 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.424509] env[63197]: DEBUG nova.compute.manager [req-87a8e002-0e15-47f7-83c8-070653b2de27 req-397a9aa8-ce77-480e-8916-6d3d9031b11a service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Received event network-vif-deleted-f3255d63-3c6b-4d3e-9134-4ab33d8da179 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1166.424710] env[63197]: INFO nova.compute.manager [req-87a8e002-0e15-47f7-83c8-070653b2de27 req-397a9aa8-ce77-480e-8916-6d3d9031b11a service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Neutron deleted interface f3255d63-3c6b-4d3e-9134-4ab33d8da179; detaching it from the instance and deleting it from the info cache [ 1166.424925] env[63197]: DEBUG nova.network.neutron [req-87a8e002-0e15-47f7-83c8-070653b2de27 req-397a9aa8-ce77-480e-8916-6d3d9031b11a service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.437257] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] The volume has not been displaced from its original location: [datastore1] volume-11297d4a-4378-4839-9470-e6876cf68883/volume-11297d4a-4378-4839-9470-e6876cf68883.vmdk. No consolidation needed. {{(pid=63197) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1166.442315] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1166.443460] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68fdab72-82d7-4b73-8e73-42c54f2019e9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.456763] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e631c99f-3f57-4994-8c09-8239732a55a7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.468078] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e4ee3c-2f96-4c3a-bdc9-f7b589b9d7cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.479961] env[63197]: DEBUG oslo_vmware.api [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1166.479961] env[63197]: value = "task-1364740" [ 1166.479961] env[63197]: _type = "Task" [ 1166.479961] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.488024] env[63197]: DEBUG oslo_vmware.api [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.494217] env[63197]: DEBUG nova.compute.manager [req-87a8e002-0e15-47f7-83c8-070653b2de27 req-397a9aa8-ce77-480e-8916-6d3d9031b11a service nova] [instance: 14456558-a814-45cd-9596-0c073c581b57] Detach interface failed, port_id=f3255d63-3c6b-4d3e-9134-4ab33d8da179, reason: Instance 14456558-a814-45cd-9596-0c073c581b57 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1166.596206] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1166.596580] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d4bb04c3-8867-4852-9e68-a7e86c879b3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.604366] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1166.604366] env[63197]: value = "task-1364741" [ 1166.604366] env[63197]: _type = "Task" [ 1166.604366] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.613641] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.675936] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.127s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.676134] env[63197]: INFO nova.compute.manager [None req-ce023c5b-c84c-4012-b6b6-5fff3ed05904 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Successfully reverted task state from rebuilding on failure for instance. [ 1166.922397] env[63197]: INFO nova.compute.manager [-] [instance: 14456558-a814-45cd-9596-0c073c581b57] Took 1.02 seconds to deallocate network for instance. [ 1166.990223] env[63197]: DEBUG oslo_vmware.api [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364740, 'name': ReconfigVM_Task, 'duration_secs': 0.234337} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.990517] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=63197) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1166.995938] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76cd1834-6e53-4933-8f91-602f2941ab9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.012075] env[63197]: DEBUG oslo_vmware.api [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1167.012075] env[63197]: value = "task-1364742" [ 1167.012075] env[63197]: _type = "Task" [ 1167.012075] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.020286] env[63197]: DEBUG oslo_vmware.api [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364742, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.115365] env[63197]: DEBUG oslo_vmware.api [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364741, 'name': PowerOnVM_Task, 'duration_secs': 0.405519} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.115974] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1167.116156] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-8f554619-c86d-4e84-b439-28f043fc79a7 tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance 'fae5ada6-f3ed-4428-8ea8-979ca98d2e0e' progress to 100 {{(pid=63197) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1167.469772] env[63197]: INFO nova.compute.manager [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Took 0.55 seconds to detach 1 volumes for instance. [ 1167.471820] env[63197]: DEBUG nova.compute.manager [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] [instance: 14456558-a814-45cd-9596-0c073c581b57] Deleting volume: 5960526b-c5d9-4c8b-ad0b-ba59179fa82a {{(pid=63197) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1167.520753] env[63197]: DEBUG oslo_vmware.api [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364742, 'name': ReconfigVM_Task, 'duration_secs': 0.131341} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.521090] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-290432', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'name': 'volume-11297d4a-4378-4839-9470-e6876cf68883', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0dc57420-b13f-4305-ae1f-6ebf61f356f3', 'attached_at': '', 'detached_at': '', 'volume_id': '11297d4a-4378-4839-9470-e6876cf68883', 'serial': '11297d4a-4378-4839-9470-e6876cf68883'} {{(pid=63197) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1168.017189] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.017492] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.017726] env[63197]: DEBUG nova.objects.instance [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lazy-loading 'resources' on Instance uuid 14456558-a814-45cd-9596-0c073c581b57 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.067639] env[63197]: DEBUG nova.objects.instance [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'flavor' on Instance uuid 0dc57420-b13f-4305-ae1f-6ebf61f356f3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.578573] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9a7209-0854-4d54-a4d2-4949d6d37d6f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.587499] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e33264-55b4-423e-ab7e-abcb182efcfe {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.616677] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff54f49b-3179-4ca6-beae-9a480fd5899f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.623401] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffced883-212a-4a3f-bce9-c8b5af198a3d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.642348] env[63197]: DEBUG nova.compute.provider_tree [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1169.016750] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.017064] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.017350] env[63197]: DEBUG nova.compute.manager [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Going to confirm migration 3 {{(pid=63197) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1169.073907] env[63197]: DEBUG oslo_concurrency.lockutils [None req-bd466f72-339d-48e1-a9d3-eb0d72b68888 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.254s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.145533] env[63197]: DEBUG nova.scheduler.client.report [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1169.563784] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1169.563992] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquired lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.564190] env[63197]: DEBUG nova.network.neutron [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1169.564379] env[63197]: DEBUG nova.objects.instance [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'info_cache' on Instance uuid fae5ada6-f3ed-4428-8ea8-979ca98d2e0e {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.650588] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.088605] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.088888] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.089165] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.089324] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.089499] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.091447] env[63197]: INFO nova.compute.manager [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Terminating instance [ 1170.093600] env[63197]: DEBUG nova.compute.manager [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1170.093666] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1170.094521] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc96362c-842f-4573-92e3-dd01cbd864cc {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.102975] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1170.103174] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-286e2fe8-d967-4164-b1a3-0b8e1e57af9e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.110115] env[63197]: DEBUG oslo_vmware.api [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1170.110115] env[63197]: value = "task-1364744" [ 1170.110115] env[63197]: _type = "Task" [ 1170.110115] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.127701] env[63197]: DEBUG oslo_vmware.api [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.170814] env[63197]: DEBUG oslo_concurrency.lockutils [None req-082eaadb-516a-4c76-ae0a-ebde79f1fd96 tempest-ServerActionsV293TestJSON-692905921 tempest-ServerActionsV293TestJSON-692905921-project-member] Lock "14456558-a814-45cd-9596-0c073c581b57" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.342s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.620078] env[63197]: DEBUG oslo_vmware.api [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364744, 'name': PowerOffVM_Task, 'duration_secs': 0.186613} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.621304] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1170.621304] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1170.621304] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1bd168a2-1783-4d6f-b101-3e8adac32dd8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.676926] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1170.677345] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Deleting contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.677612] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleting the datastore file [datastore1] 0dc57420-b13f-4305-ae1f-6ebf61f356f3 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.677899] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6260643a-96b7-45f1-831f-c56a9d82fb4e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.684695] env[63197]: DEBUG oslo_vmware.api [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for the task: (returnval){ [ 1170.684695] env[63197]: value = "task-1364746" [ 1170.684695] env[63197]: _type = "Task" [ 1170.684695] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.692683] env[63197]: DEBUG oslo_vmware.api [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.859279] env[63197]: DEBUG nova.network.neutron [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance_info_cache with network_info: [{"id": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "address": "fa:16:3e:b4:f5:5e", "network": {"id": "1a677f3d-c335-4533-9ff9-cd4b37db0a6d", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-849806371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0ceb1016d6d34bff8880dca42d495377", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3cecd3aa-c7", "ovs_interfaceid": "3cecd3aa-c700-408e-b8ce-50c12cc106a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.982870] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1170.983381] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63197) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1171.195583] env[63197]: DEBUG oslo_vmware.api [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Task: {'id': task-1364746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179268} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.195583] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.195764] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Deleted contents of the VM from datastore datastore1 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1171.195901] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1171.196102] env[63197]: INFO nova.compute.manager [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1171.196338] env[63197]: DEBUG oslo.service.loopingcall [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.196529] env[63197]: DEBUG nova.compute.manager [-] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1171.196613] env[63197]: DEBUG nova.network.neutron [-] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1171.362280] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Releasing lock "refresh_cache-fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.362920] env[63197]: DEBUG nova.objects.instance [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'migration_context' on Instance uuid fae5ada6-f3ed-4428-8ea8-979ca98d2e0e {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.478239] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.478239] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.874426] env[63197]: DEBUG nova.objects.base [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63197) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1171.874948] env[63197]: DEBUG nova.compute.manager [req-29fdec58-cdcb-41a2-b832-f3571c84b28a req-94a43620-bbba-490f-a328-72c1f8bf9e82 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Received event network-vif-deleted-7b0c52af-a5fc-4266-8dd0-73c379e73cdd {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1171.874948] env[63197]: INFO nova.compute.manager [req-29fdec58-cdcb-41a2-b832-f3571c84b28a req-94a43620-bbba-490f-a328-72c1f8bf9e82 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Neutron deleted interface 7b0c52af-a5fc-4266-8dd0-73c379e73cdd; detaching it from the instance and deleting it from the info cache [ 1171.875037] env[63197]: DEBUG nova.network.neutron [req-29fdec58-cdcb-41a2-b832-f3571c84b28a req-94a43620-bbba-490f-a328-72c1f8bf9e82 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.877424] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b4de61-a851-4336-b398-d933db33983d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.896999] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e57adb01-e474-4c2a-b77c-21c71a2a4169 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.902983] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1171.902983] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5242218b-d7be-209c-b122-6c56489a2252" [ 1171.902983] env[63197]: _type = "Task" [ 1171.902983] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.911236] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5242218b-d7be-209c-b122-6c56489a2252, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.979582] env[63197]: DEBUG nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Starting instance... {{(pid=63197) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1172.322085] env[63197]: DEBUG nova.network.neutron [-] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.382017] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9632745f-e416-4b56-9b51-05f9233fa0ef {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.390923] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a086fa3-7367-46e6-9cdf-b8d6af26dab3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.412351] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5242218b-d7be-209c-b122-6c56489a2252, 'name': SearchDatastore_Task, 'duration_secs': 0.007439} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.418063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.418324] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.422549] env[63197]: DEBUG nova.compute.manager [req-29fdec58-cdcb-41a2-b832-f3571c84b28a req-94a43620-bbba-490f-a328-72c1f8bf9e82 service nova] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Detach interface failed, port_id=7b0c52af-a5fc-4266-8dd0-73c379e73cdd, reason: Instance 0dc57420-b13f-4305-ae1f-6ebf61f356f3 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1172.502912] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.826662] env[63197]: INFO nova.compute.manager [-] [instance: 0dc57420-b13f-4305-ae1f-6ebf61f356f3] Took 1.63 seconds to deallocate network for instance. [ 1172.984104] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1172.996126] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06909c8f-4f84-4f37-a830-bc5ae3d7f921 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.004278] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83ea309-33a3-4b12-805d-7cbf2a4d873c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.033564] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb634f1-64c9-47eb-9b6c-dcc3f9a78578 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.040403] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a335d8-4dc3-4b4b-9e1e-8d8ffaadd4ab {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.052964] env[63197]: DEBUG nova.compute.provider_tree [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.332861] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.556552] env[63197]: DEBUG nova.scheduler.client.report [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1173.980904] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1173.982549] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1174.567279] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.149s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.570113] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.067s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.571890] env[63197]: INFO nova.compute.claims [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1174.983513] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1175.125087] env[63197]: INFO nova.scheduler.client.report [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted allocation for migration b65c0df8-5de5-42d3-80b1-146ced5d985e [ 1175.630211] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.613s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.635742] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817a7550-7239-4886-93e8-54e0e12b102a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.642988] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ef1c47-1688-4966-8a27-7dbbbed0fb96 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.672936] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44e0b5e-7b55-4f50-91fb-5b18e3e19494 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.679796] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90daa9a9-fc05-47dc-99c0-784193734c02 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.692683] env[63197]: DEBUG nova.compute.provider_tree [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.983671] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1176.196014] env[63197]: DEBUG nova.scheduler.client.report [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.700309] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.130s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.700824] env[63197]: DEBUG nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Start building networks asynchronously for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1176.703399] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.371s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.703621] env[63197]: DEBUG nova.objects.instance [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lazy-loading 'resources' on Instance uuid 0dc57420-b13f-4305-ae1f-6ebf61f356f3 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1176.983196] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager.update_available_resource {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1177.043144] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.043422] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.043640] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.043825] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.043997] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.045925] env[63197]: INFO nova.compute.manager [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Terminating instance [ 1177.047738] env[63197]: DEBUG nova.compute.manager [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1177.047936] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1177.048778] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3aebb6-e3db-42fe-b21d-115be1e24e22 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.056113] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1177.056341] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9f0fe9a-06d2-4f83-951e-618f7544156f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.063391] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1177.063391] env[63197]: value = "task-1364747" [ 1177.063391] env[63197]: _type = "Task" [ 1177.063391] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.070648] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364747, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.206954] env[63197]: DEBUG nova.compute.utils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Using /dev/sd instead of None {{(pid=63197) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1177.211193] env[63197]: DEBUG nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Allocating IP information in the background. {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1177.211367] env[63197]: DEBUG nova.network.neutron [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] allocate_for_instance() {{(pid=63197) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1177.250756] env[63197]: DEBUG nova.policy [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1787d96096e1445cb6f054fff88d9e4c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0632089ea0a242a184263c8c974db8a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63197) authorize /opt/stack/nova/nova/policy.py:201}} [ 1177.273208] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c359239-3966-4589-a2c6-3c473f273650 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.280742] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b718bb-d9ca-4b72-b025-5964b26ea28f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.310032] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96702574-7c0d-4562-861c-afd01ba74717 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.317125] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8b62e8-3d5f-4206-9448-3350f37a42bb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.331368] env[63197]: DEBUG nova.compute.provider_tree [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1177.486090] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.506319] env[63197]: DEBUG nova.network.neutron [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Successfully created port: 082337d6-8fd2-4a11-b908-e9e9d00ca591 {{(pid=63197) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1177.573428] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364747, 'name': PowerOffVM_Task, 'duration_secs': 0.184439} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.573703] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1177.573875] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1177.574137] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dffef214-4cdc-4dee-ac93-35cd2723653c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.635489] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1177.635744] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1177.635921] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleting the datastore file [datastore2] fae5ada6-f3ed-4428-8ea8-979ca98d2e0e {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1177.636272] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09f40209-2c17-4c7d-8a21-2a4545fc8d2b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.642214] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for the task: (returnval){ [ 1177.642214] env[63197]: value = "task-1364749" [ 1177.642214] env[63197]: _type = "Task" [ 1177.642214] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.649848] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.711729] env[63197]: DEBUG nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Start building block device mappings for instance. {{(pid=63197) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1177.835404] env[63197]: DEBUG nova.scheduler.client.report [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1178.152255] env[63197]: DEBUG oslo_vmware.api [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Task: {'id': task-1364749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137425} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.152517] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1178.152704] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1178.152924] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1178.153123] env[63197]: INFO nova.compute.manager [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1178.153369] env[63197]: DEBUG oslo.service.loopingcall [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1178.153563] env[63197]: DEBUG nova.compute.manager [-] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1178.153659] env[63197]: DEBUG nova.network.neutron [-] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1178.340319] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.342520] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.857s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.343415] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.343529] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63197) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1178.344430] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a2b46f-ea42-4292-aa00-46881be61ba1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.354191] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6368091b-2238-4412-b06d-1bceb5ea8ea2 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.369033] env[63197]: INFO nova.scheduler.client.report [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Deleted allocations for instance 0dc57420-b13f-4305-ae1f-6ebf61f356f3 [ 1178.371153] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719d05db-bcc2-41d5-aa4a-7e444a0692e8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.382450] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb2ca8d-2726-4409-8566-f59ec1cd881b {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.386500] env[63197]: DEBUG nova.compute.manager [req-03a83dd3-ddf5-4680-bc59-0dd4b4fc10a9 req-94573a10-c04f-419e-9d6e-3b6ee6e886d3 service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Received event network-vif-deleted-3cecd3aa-c700-408e-b8ce-50c12cc106a0 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1178.386697] env[63197]: INFO nova.compute.manager [req-03a83dd3-ddf5-4680-bc59-0dd4b4fc10a9 req-94573a10-c04f-419e-9d6e-3b6ee6e886d3 service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Neutron deleted interface 3cecd3aa-c700-408e-b8ce-50c12cc106a0; detaching it from the instance and deleting it from the info cache [ 1178.386868] env[63197]: DEBUG nova.network.neutron [req-03a83dd3-ddf5-4680-bc59-0dd4b4fc10a9 req-94573a10-c04f-419e-9d6e-3b6ee6e886d3 service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.417115] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181462MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=63197) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1178.417290] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.417469] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.722164] env[63197]: DEBUG nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Start spawning the instance on the hypervisor. {{(pid=63197) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1178.747557] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-25T09:42:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-25T09:41:52Z,direct_url=,disk_format='vmdk',id=f760fedc-0b5b-4c56-acbe-239b47c945e8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='ba0c482e322a4865bd946d9a59e634dc',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-25T09:41:53Z,virtual_size=,visibility=), allow threads: False {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1178.747820] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Flavor limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1178.747979] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Image limits 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1178.748177] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Flavor pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1178.748326] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Image pref 0:0:0 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1178.748473] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63197) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1178.748679] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1178.748840] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1178.749016] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Got 1 possible topologies {{(pid=63197) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1178.749197] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1178.749371] env[63197]: DEBUG nova.virt.hardware [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63197) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1178.750237] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1070fd3b-dddd-4bc6-b150-04d4fd706b66 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.757878] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf1f6c3-63ea-49e4-8211-09d023184878 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.861987] env[63197]: DEBUG nova.network.neutron [-] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.879894] env[63197]: DEBUG oslo_concurrency.lockutils [None req-5e8641f2-189f-485a-8b0f-0703a62e18b3 tempest-AttachVolumeNegativeTest-617765548 tempest-AttachVolumeNegativeTest-617765548-project-member] Lock "0dc57420-b13f-4305-ae1f-6ebf61f356f3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.791s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.886721] env[63197]: DEBUG nova.compute.manager [req-26e81a76-fe1f-4775-a95c-4aa25ab558a1 req-195f3dfd-3e64-41a9-a6b5-e3e6b8719e76 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Received event network-vif-plugged-082337d6-8fd2-4a11-b908-e9e9d00ca591 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1178.886721] env[63197]: DEBUG oslo_concurrency.lockutils [req-26e81a76-fe1f-4775-a95c-4aa25ab558a1 req-195f3dfd-3e64-41a9-a6b5-e3e6b8719e76 service nova] Acquiring lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.887041] env[63197]: DEBUG oslo_concurrency.lockutils [req-26e81a76-fe1f-4775-a95c-4aa25ab558a1 req-195f3dfd-3e64-41a9-a6b5-e3e6b8719e76 service nova] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.887213] env[63197]: DEBUG oslo_concurrency.lockutils [req-26e81a76-fe1f-4775-a95c-4aa25ab558a1 req-195f3dfd-3e64-41a9-a6b5-e3e6b8719e76 service nova] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.887387] env[63197]: DEBUG nova.compute.manager [req-26e81a76-fe1f-4775-a95c-4aa25ab558a1 req-195f3dfd-3e64-41a9-a6b5-e3e6b8719e76 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] No waiting events found dispatching network-vif-plugged-082337d6-8fd2-4a11-b908-e9e9d00ca591 {{(pid=63197) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1178.887582] env[63197]: WARNING nova.compute.manager [req-26e81a76-fe1f-4775-a95c-4aa25ab558a1 req-195f3dfd-3e64-41a9-a6b5-e3e6b8719e76 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Received unexpected event network-vif-plugged-082337d6-8fd2-4a11-b908-e9e9d00ca591 for instance with vm_state building and task_state spawning. [ 1178.889429] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0010818-a989-47c9-8162-4dc07d46085d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.899675] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0776fe-8e26-4035-93a3-dd3614d9f47e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.926013] env[63197]: DEBUG nova.compute.manager [req-03a83dd3-ddf5-4680-bc59-0dd4b4fc10a9 req-94573a10-c04f-419e-9d6e-3b6ee6e886d3 service nova] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Detach interface failed, port_id=3cecd3aa-c700-408e-b8ce-50c12cc106a0, reason: Instance fae5ada6-f3ed-4428-8ea8-979ca98d2e0e could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1179.363724] env[63197]: INFO nova.compute.manager [-] [instance: fae5ada6-f3ed-4428-8ea8-979ca98d2e0e] Took 1.21 seconds to deallocate network for instance. [ 1179.446329] env[63197]: DEBUG nova.network.neutron [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Successfully updated port: 082337d6-8fd2-4a11-b908-e9e9d00ca591 {{(pid=63197) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1179.450837] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance ad216370-3f6d-4ebe-a956-7dd3bf05511f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1179.450837] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance fae5ada6-f3ed-4428-8ea8-979ca98d2e0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1179.450837] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Instance 87acbd7c-2a03-439b-ab06-81e6f4b8d903 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63197) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1179.450837] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1179.450837] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63197) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1179.503638] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8ef804-4e3a-41df-bc66-f3ccdcdb91d4 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.511612] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911f2846-9be4-445a-9bb7-ca98e7ccf5b7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.542627] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f4e881-e91e-4fb6-828f-090863240384 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.550220] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2715607-6859-48ae-9050-414c9ce14d5d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.563477] env[63197]: DEBUG nova.compute.provider_tree [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1179.871904] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1179.949340] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "refresh_cache-87acbd7c-2a03-439b-ab06-81e6f4b8d903" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.949495] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquired lock "refresh_cache-87acbd7c-2a03-439b-ab06-81e6f4b8d903" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.949648] env[63197]: DEBUG nova.network.neutron [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Building network info cache for instance {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1180.066760] env[63197]: DEBUG nova.scheduler.client.report [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1180.413394] env[63197]: DEBUG nova.compute.manager [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Received event network-changed-082337d6-8fd2-4a11-b908-e9e9d00ca591 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1180.413596] env[63197]: DEBUG nova.compute.manager [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Refreshing instance network info cache due to event network-changed-082337d6-8fd2-4a11-b908-e9e9d00ca591. {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1180.413785] env[63197]: DEBUG oslo_concurrency.lockutils [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] Acquiring lock "refresh_cache-87acbd7c-2a03-439b-ab06-81e6f4b8d903" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1180.500885] env[63197]: DEBUG nova.network.neutron [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Instance cache missing network info. {{(pid=63197) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1180.572039] env[63197]: DEBUG nova.compute.resource_tracker [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63197) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1180.572245] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.155s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.572519] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.701s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.572744] env[63197]: DEBUG nova.objects.instance [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lazy-loading 'resources' on Instance uuid fae5ada6-f3ed-4428-8ea8-979ca98d2e0e {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1180.645035] env[63197]: DEBUG nova.network.neutron [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Updating instance_info_cache with network_info: [{"id": "082337d6-8fd2-4a11-b908-e9e9d00ca591", "address": "fa:16:3e:83:37:3a", "network": {"id": "77171661-3976-4c73-8bb6-8955bfa10dd6", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1381724749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0632089ea0a242a184263c8c974db8a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap082337d6-8f", "ovs_interfaceid": "082337d6-8fd2-4a11-b908-e9e9d00ca591", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.128052] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4eab085-7daf-4ef6-8ace-f9f9fc12b8c8 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.135039] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde7a4ea-fa7f-4207-90b6-02e890d6bdb9 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.163647] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Releasing lock "refresh_cache-87acbd7c-2a03-439b-ab06-81e6f4b8d903" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1181.163961] env[63197]: DEBUG nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Instance network_info: |[{"id": "082337d6-8fd2-4a11-b908-e9e9d00ca591", "address": "fa:16:3e:83:37:3a", "network": {"id": "77171661-3976-4c73-8bb6-8955bfa10dd6", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1381724749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0632089ea0a242a184263c8c974db8a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap082337d6-8f", "ovs_interfaceid": "082337d6-8fd2-4a11-b908-e9e9d00ca591", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63197) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1181.164482] env[63197]: DEBUG oslo_concurrency.lockutils [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] Acquired lock "refresh_cache-87acbd7c-2a03-439b-ab06-81e6f4b8d903" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.164677] env[63197]: DEBUG nova.network.neutron [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Refreshing network info cache for port 082337d6-8fd2-4a11-b908-e9e9d00ca591 {{(pid=63197) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1181.166042] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:37:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3e0aae3-33d1-403b-bfaf-306f77a1422e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '082337d6-8fd2-4a11-b908-e9e9d00ca591', 'vif_model': 'vmxnet3'}] {{(pid=63197) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1181.172989] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Creating folder: Project (0632089ea0a242a184263c8c974db8a1). Parent ref: group-v290286. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1181.173702] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19150e8d-16fb-4a39-b58e-3627e3cd6df1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.176932] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a93127e7-b287-4246-aa96-1b374ffabf9f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.183858] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b9c7953-43fb-4562-bb9e-1e36035610f6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.188678] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Created folder: Project (0632089ea0a242a184263c8c974db8a1) in parent group-v290286. [ 1181.188855] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Creating folder: Instances. Parent ref: group-v290433. {{(pid=63197) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1181.189380] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4ae4c2d-5793-487a-8bea-9dfd90c1ac8e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.198596] env[63197]: DEBUG nova.compute.provider_tree [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1181.200556] env[63197]: INFO nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Created folder: Instances in parent group-v290433. [ 1181.200781] env[63197]: DEBUG oslo.service.loopingcall [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1181.201149] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Creating VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1181.201346] env[63197]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-156ab36c-ce0b-438b-a5df-810977ab97d1 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.218994] env[63197]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1181.218994] env[63197]: value = "task-1364753" [ 1181.218994] env[63197]: _type = "Task" [ 1181.218994] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.225912] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364753, 'name': CreateVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.702187] env[63197]: DEBUG nova.scheduler.client.report [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1181.729138] env[63197]: DEBUG oslo_vmware.api [-] Task: {'id': task-1364753, 'name': CreateVM_Task, 'duration_secs': 0.316661} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.729282] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Created VM on the ESX host {{(pid=63197) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1181.729902] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.730084] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.730417] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1181.730659] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9dd97f4c-1611-4902-8915-0620f88a81fb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.734820] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1181.734820] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280a362-8b8e-0c48-eba9-5089aecaa3f6" [ 1181.734820] env[63197]: _type = "Task" [ 1181.734820] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.744242] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280a362-8b8e-0c48-eba9-5089aecaa3f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.864022] env[63197]: DEBUG nova.network.neutron [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Updated VIF entry in instance network info cache for port 082337d6-8fd2-4a11-b908-e9e9d00ca591. {{(pid=63197) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1181.864560] env[63197]: DEBUG nova.network.neutron [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Updating instance_info_cache with network_info: [{"id": "082337d6-8fd2-4a11-b908-e9e9d00ca591", "address": "fa:16:3e:83:37:3a", "network": {"id": "77171661-3976-4c73-8bb6-8955bfa10dd6", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1381724749-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0632089ea0a242a184263c8c974db8a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3e0aae3-33d1-403b-bfaf-306f77a1422e", "external-id": "nsx-vlan-transportzone-211", "segmentation_id": 211, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap082337d6-8f", "ovs_interfaceid": "082337d6-8fd2-4a11-b908-e9e9d00ca591", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.208999] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.636s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.228229] env[63197]: INFO nova.scheduler.client.report [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Deleted allocations for instance fae5ada6-f3ed-4428-8ea8-979ca98d2e0e [ 1182.244694] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]5280a362-8b8e-0c48-eba9-5089aecaa3f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009308} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.244694] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.244839] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Processing image f760fedc-0b5b-4c56-acbe-239b47c945e8 {{(pid=63197) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1182.245063] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.245229] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.245411] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.245667] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a18590fd-9e0e-4332-bf1a-f0d7e804f17d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.256333] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63197) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.256509] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63197) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1182.257198] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1a6d238-db6f-47f6-a11a-bdbca6db2c39 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.261824] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1182.261824] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52279069-b699-6076-03e8-600c4ceb0cd8" [ 1182.261824] env[63197]: _type = "Task" [ 1182.261824] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.269031] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52279069-b699-6076-03e8-600c4ceb0cd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.367391] env[63197]: DEBUG oslo_concurrency.lockutils [req-8f504a99-71cf-4b41-ade1-3b940aa2a192 req-e1c034c5-ead5-49de-88df-cdb4d94d8e00 service nova] Releasing lock "refresh_cache-87acbd7c-2a03-439b-ab06-81e6f4b8d903" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.736130] env[63197]: DEBUG oslo_concurrency.lockutils [None req-2b4c2f4d-9952-437f-aba6-bd1ba65e5f1a tempest-DeleteServersTestJSON-503485045 tempest-DeleteServersTestJSON-503485045-project-member] Lock "fae5ada6-f3ed-4428-8ea8-979ca98d2e0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.693s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1182.771818] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]52279069-b699-6076-03e8-600c4ceb0cd8, 'name': SearchDatastore_Task, 'duration_secs': 0.008256} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.772602] env[63197]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc8f59fd-8636-4d6b-889e-3a982a2a3a76 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.777653] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1182.777653] env[63197]: value = "session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529b6f1c-0f39-b99f-70eb-74c5bcfe8822" [ 1182.777653] env[63197]: _type = "Task" [ 1182.777653] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.785040] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529b6f1c-0f39-b99f-70eb-74c5bcfe8822, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.288090] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': session[52c0eeb9-0faa-d450-cc75-7aba0ba2703e]529b6f1c-0f39-b99f-70eb-74c5bcfe8822, 'name': SearchDatastore_Task, 'duration_secs': 0.008768} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.288341] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk" {{(pid=63197) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.288584] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 87acbd7c-2a03-439b-ab06-81e6f4b8d903/87acbd7c-2a03-439b-ab06-81e6f4b8d903.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1183.288854] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36e674fd-b604-4c19-809c-277589affc80 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.295298] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1183.295298] env[63197]: value = "task-1364755" [ 1183.295298] env[63197]: _type = "Task" [ 1183.295298] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.303425] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364755, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.573700] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1183.574021] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Starting heal instance info cache {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1183.805444] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364755, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451496} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.805673] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f760fedc-0b5b-4c56-acbe-239b47c945e8/f760fedc-0b5b-4c56-acbe-239b47c945e8.vmdk to [datastore2] 87acbd7c-2a03-439b-ab06-81e6f4b8d903/87acbd7c-2a03-439b-ab06-81e6f4b8d903.vmdk {{(pid=63197) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1183.805841] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Extending root virtual disk to 1048576 {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1183.806116] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6084b23c-583a-4c9b-b011-14ba92051603 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.811502] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1183.811502] env[63197]: value = "task-1364757" [ 1183.811502] env[63197]: _type = "Task" [ 1183.811502] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.820271] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364757, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.321217] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364757, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060022} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.321217] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Extended root virtual disk {{(pid=63197) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1184.321881] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc067c0e-05d0-4cde-950c-b390878ad31a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.343215] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 87acbd7c-2a03-439b-ab06-81e6f4b8d903/87acbd7c-2a03-439b-ab06-81e6f4b8d903.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1184.343479] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-695fc4e4-63e8-404c-9217-3cbda2203ba3 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.362933] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1184.362933] env[63197]: value = "task-1364758" [ 1184.362933] env[63197]: _type = "Task" [ 1184.362933] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.370491] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364758, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.579758] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Didn't find any instances for network info cache update. {{(pid=63197) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1184.579980] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.580171] env[63197]: DEBUG oslo_service.periodic_task [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Running periodic task ComputeManager._sync_power_states {{(pid=63197) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1184.873521] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364758, 'name': ReconfigVM_Task, 'duration_secs': 0.275616} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.873521] env[63197]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 87acbd7c-2a03-439b-ab06-81e6f4b8d903/87acbd7c-2a03-439b-ab06-81e6f4b8d903.vmdk or device None with type sparse {{(pid=63197) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1184.874025] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf698bdf-3631-4a63-977f-6804786d459d {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.880097] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1184.880097] env[63197]: value = "task-1364759" [ 1184.880097] env[63197]: _type = "Task" [ 1184.880097] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.887321] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364759, 'name': Rename_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.060194] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.060490] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.060708] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.060899] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.061095] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.063253] env[63197]: INFO nova.compute.manager [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Terminating instance [ 1185.065117] env[63197]: DEBUG nova.compute.manager [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1185.065328] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1185.066191] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc95b331-fb11-4159-bb90-01edcf353f4f {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.073713] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.074298] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce8701e2-1d4a-42a2-8358-7d269fedeca6 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.080132] env[63197]: DEBUG oslo_vmware.api [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1185.080132] env[63197]: value = "task-1364760" [ 1185.080132] env[63197]: _type = "Task" [ 1185.080132] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.083668] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Getting list of instances from cluster (obj){ [ 1185.083668] env[63197]: value = "domain-c8" [ 1185.083668] env[63197]: _type = "ClusterComputeResource" [ 1185.083668] env[63197]: } {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1185.084531] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5906d9e5-9893-4a40-9057-2a73d8896953 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.093473] env[63197]: DEBUG oslo_vmware.api [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.096819] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Got total of 2 instances {{(pid=63197) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1185.096974] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1185.097177] env[63197]: DEBUG nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Triggering sync for uuid 87acbd7c-2a03-439b-ab06-81e6f4b8d903 {{(pid=63197) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1185.097483] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.097709] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Acquiring lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.390761] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364759, 'name': Rename_Task} progress is 14%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.591102] env[63197]: DEBUG oslo_vmware.api [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364760, 'name': PowerOffVM_Task, 'duration_secs': 0.232578} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.591102] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1185.591287] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1185.593739] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6333177-ab2d-4b4a-b0c9-a8e185eef37c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.892406] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364759, 'name': Rename_Task, 'duration_secs': 0.846855} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.892778] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Powering on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1185.893051] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8d8e6f0-9531-42d5-8db0-36b615597176 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.898901] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1185.898901] env[63197]: value = "task-1364762" [ 1185.898901] env[63197]: _type = "Task" [ 1185.898901] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.905794] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364762, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.978938] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1185.979220] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1185.979411] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleting the datastore file [datastore2] ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1185.979721] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c54804b-2e80-497e-908e-10111b040b19 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.987174] env[63197]: DEBUG oslo_vmware.api [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for the task: (returnval){ [ 1185.987174] env[63197]: value = "task-1364763" [ 1185.987174] env[63197]: _type = "Task" [ 1185.987174] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.997415] env[63197]: DEBUG oslo_vmware.api [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364763, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.408481] env[63197]: DEBUG oslo_vmware.api [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364762, 'name': PowerOnVM_Task, 'duration_secs': 0.472299} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.408769] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Powered on the VM {{(pid=63197) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1186.408947] env[63197]: INFO nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1186.409338] env[63197]: DEBUG nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Checking state {{(pid=63197) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.410099] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e663a83a-a223-4c30-ad55-51f401aa43cb {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.496839] env[63197]: DEBUG oslo_vmware.api [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364763, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.929544] env[63197]: INFO nova.compute.manager [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Took 14.44 seconds to build instance. [ 1186.997338] env[63197]: DEBUG oslo_vmware.api [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Task: {'id': task-1364763, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.568184} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.997604] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1186.997787] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1186.997963] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.998151] env[63197]: INFO nova.compute.manager [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Took 1.93 seconds to destroy the instance on the hypervisor. [ 1186.998397] env[63197]: DEBUG oslo.service.loopingcall [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1186.998593] env[63197]: DEBUG nova.compute.manager [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1186.998688] env[63197]: DEBUG nova.network.neutron [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1187.416560] env[63197]: DEBUG nova.compute.manager [req-ab928064-0cd3-4d1f-a377-e84218c57207 req-f65ce9d0-e64e-47aa-bc47-5c5d91f92634 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Received event network-vif-deleted-ec9b6a1b-f520-4f03-b0ce-51ee8346b1af {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1187.416914] env[63197]: INFO nova.compute.manager [req-ab928064-0cd3-4d1f-a377-e84218c57207 req-f65ce9d0-e64e-47aa-bc47-5c5d91f92634 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Neutron deleted interface ec9b6a1b-f520-4f03-b0ce-51ee8346b1af; detaching it from the instance and deleting it from the info cache [ 1187.416988] env[63197]: DEBUG nova.network.neutron [req-ab928064-0cd3-4d1f-a377-e84218c57207 req-f65ce9d0-e64e-47aa-bc47-5c5d91f92634 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.430641] env[63197]: DEBUG oslo_concurrency.lockutils [None req-ec3b3526-9ca7-42f1-8333-dd1c20c81729 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.953s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.430905] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.333s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1187.431102] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] During sync_power_state the instance has a pending task (spawning). Skip. [ 1187.431284] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.889512] env[63197]: DEBUG nova.network.neutron [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.921412] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f0bb1545-c396-4f75-86d2-70964885876e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.932168] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9552680e-65e9-4b4f-b146-303c2ded2ce7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.957653] env[63197]: DEBUG nova.compute.manager [req-ab928064-0cd3-4d1f-a377-e84218c57207 req-f65ce9d0-e64e-47aa-bc47-5c5d91f92634 service nova] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Detach interface failed, port_id=ec9b6a1b-f520-4f03-b0ce-51ee8346b1af, reason: Instance ad216370-3f6d-4ebe-a956-7dd3bf05511f could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1188.392190] env[63197]: INFO nova.compute.manager [-] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] Took 1.39 seconds to deallocate network for instance. [ 1188.898954] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.899310] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.899501] env[63197]: DEBUG nova.objects.instance [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lazy-loading 'resources' on Instance uuid ad216370-3f6d-4ebe-a956-7dd3bf05511f {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.919454] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.919705] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.919913] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.920123] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.920321] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.922309] env[63197]: INFO nova.compute.manager [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Terminating instance [ 1188.924129] env[63197]: DEBUG nova.compute.manager [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Start destroying the instance on the hypervisor. {{(pid=63197) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1188.924331] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Destroying instance {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1188.925220] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496e8e68-083b-435c-9b7b-152545504b68 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.933231] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Powering off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.933476] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02dd4e75-69d3-4038-9753-60dae7a2f431 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.939774] env[63197]: DEBUG oslo_vmware.api [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1188.939774] env[63197]: value = "task-1364765" [ 1188.939774] env[63197]: _type = "Task" [ 1188.939774] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.947469] env[63197]: DEBUG oslo_vmware.api [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364765, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.444801] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb7c790-c0a3-47bb-a473-e0e2e0d8af1e {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.452538] env[63197]: DEBUG oslo_vmware.api [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364765, 'name': PowerOffVM_Task, 'duration_secs': 0.197468} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.454375] env[63197]: DEBUG nova.virt.vmwareapi.vm_util [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Powered off the VM {{(pid=63197) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1189.454553] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Unregistering the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1189.454826] env[63197]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77dcc4c7-a16f-4f17-9cbc-b831acc8b5ee {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.456868] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9094f866-70fb-47b3-932d-e082a20e8f7c {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.486323] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b9f3be-ba0d-439e-b59f-a42706248551 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.493516] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e437e50-7e12-4c4a-8c99-17399dd81597 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.507569] env[63197]: DEBUG nova.compute.provider_tree [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.517193] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Unregistered the VM {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1189.517417] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Deleting contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1189.517598] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Deleting the datastore file [datastore2] 87acbd7c-2a03-439b-ab06-81e6f4b8d903 {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1189.518452] env[63197]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c1dbdd0-058f-44f9-a59b-6d52cf2697e7 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.524226] env[63197]: DEBUG oslo_vmware.api [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for the task: (returnval){ [ 1189.524226] env[63197]: value = "task-1364767" [ 1189.524226] env[63197]: _type = "Task" [ 1189.524226] env[63197]: } to complete. {{(pid=63197) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.531802] env[63197]: DEBUG oslo_vmware.api [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.011436] env[63197]: DEBUG nova.scheduler.client.report [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.034232] env[63197]: DEBUG oslo_vmware.api [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Task: {'id': task-1364767, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127654} completed successfully. {{(pid=63197) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.034560] env[63197]: DEBUG nova.virt.vmwareapi.ds_util [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Deleted the datastore file {{(pid=63197) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1190.034755] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Deleted contents of the VM from datastore datastore2 {{(pid=63197) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1190.034937] env[63197]: DEBUG nova.virt.vmwareapi.vmops [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Instance destroyed {{(pid=63197) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1190.035152] env[63197]: INFO nova.compute.manager [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1190.035398] env[63197]: DEBUG oslo.service.loopingcall [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63197) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1190.035585] env[63197]: DEBUG nova.compute.manager [-] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Deallocating network for instance {{(pid=63197) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1190.035678] env[63197]: DEBUG nova.network.neutron [-] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] deallocate_for_instance() {{(pid=63197) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1190.281323] env[63197]: DEBUG nova.compute.manager [req-34651503-8e50-40f5-b02d-c8e724d81bdf req-d5505a03-8070-4417-b650-c74f7cfe757b service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Received event network-vif-deleted-082337d6-8fd2-4a11-b908-e9e9d00ca591 {{(pid=63197) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1190.281527] env[63197]: INFO nova.compute.manager [req-34651503-8e50-40f5-b02d-c8e724d81bdf req-d5505a03-8070-4417-b650-c74f7cfe757b service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Neutron deleted interface 082337d6-8fd2-4a11-b908-e9e9d00ca591; detaching it from the instance and deleting it from the info cache [ 1190.281706] env[63197]: DEBUG nova.network.neutron [req-34651503-8e50-40f5-b02d-c8e724d81bdf req-d5505a03-8070-4417-b650-c74f7cfe757b service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.516672] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.543976] env[63197]: INFO nova.scheduler.client.report [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Deleted allocations for instance ad216370-3f6d-4ebe-a956-7dd3bf05511f [ 1190.754567] env[63197]: DEBUG nova.network.neutron [-] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Updating instance_info_cache with network_info: [] {{(pid=63197) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.784658] env[63197]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-058eda27-b105-484e-b381-5007ecfa13de {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.794580] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae333ec-4487-4071-95ff-8a979a0e9915 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.816708] env[63197]: DEBUG nova.compute.manager [req-34651503-8e50-40f5-b02d-c8e724d81bdf req-d5505a03-8070-4417-b650-c74f7cfe757b service nova] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Detach interface failed, port_id=082337d6-8fd2-4a11-b908-e9e9d00ca591, reason: Instance 87acbd7c-2a03-439b-ab06-81e6f4b8d903 could not be found. {{(pid=63197) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1191.053896] env[63197]: DEBUG oslo_concurrency.lockutils [None req-a33ff808-d773-4b84-9469-86041e721a54 tempest-AttachVolumeShelveTestJSON-996991588 tempest-AttachVolumeShelveTestJSON-996991588-project-member] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.993s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.054869] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.957s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.055120] env[63197]: INFO nova.compute.manager [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] [instance: ad216370-3f6d-4ebe-a956-7dd3bf05511f] During sync_power_state the instance has a pending task (deleting). Skip. [ 1191.055322] env[63197]: DEBUG oslo_concurrency.lockutils [None req-21c587fe-9602-44c7-9fea-0815bf3544cf None None] Lock "ad216370-3f6d-4ebe-a956-7dd3bf05511f" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.256607] env[63197]: INFO nova.compute.manager [-] [instance: 87acbd7c-2a03-439b-ab06-81e6f4b8d903] Took 1.22 seconds to deallocate network for instance. [ 1191.764567] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.764924] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.765313] env[63197]: DEBUG nova.objects.instance [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lazy-loading 'resources' on Instance uuid 87acbd7c-2a03-439b-ab06-81e6f4b8d903 {{(pid=63197) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.301054] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f391fe8c-5c77-4371-a4a7-1ca6fad0c73a {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.308890] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6769192d-a491-476c-a258-0c3ac53e5dca {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.337562] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd411b9a-961c-4677-8011-d7d9ea82eb54 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.344437] env[63197]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439fd48a-d436-4c29-a68b-a91338603210 {{(pid=63197) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.357044] env[63197]: DEBUG nova.compute.provider_tree [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Inventory has not changed in ProviderTree for provider: e9e01824-bb93-4102-9e5b-66d5dbc1990b {{(pid=63197) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.860055] env[63197]: DEBUG nova.scheduler.client.report [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Inventory has not changed for provider e9e01824-bb93-4102-9e5b-66d5dbc1990b based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63197) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1193.365107] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.382552] env[63197]: INFO nova.scheduler.client.report [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Deleted allocations for instance 87acbd7c-2a03-439b-ab06-81e6f4b8d903 [ 1193.890079] env[63197]: DEBUG oslo_concurrency.lockutils [None req-112664f3-a986-49c1-8b81-2ee0ec3ef495 tempest-ServerMetadataTestJSON-1756718402 tempest-ServerMetadataTestJSON-1756718402-project-member] Lock "87acbd7c-2a03-439b-ab06-81e6f4b8d903" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.970s {{(pid=63197) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}